2025-12-04T09:15:26.3090635Z Current runner version: '2.330.0' 2025-12-04T09:15:26.3096171Z Runner name: 'i-016559b86632fc7da' 2025-12-04T09:15:26.3096924Z Runner group name: 'default' 2025-12-04T09:15:26.3097734Z Machine name: 'ip-10-0-66-49' 2025-12-04T09:15:26.3100445Z ##[group]GITHUB_TOKEN Permissions 2025-12-04T09:15:26.3102763Z Contents: read 2025-12-04T09:15:26.3103291Z Metadata: read 2025-12-04T09:15:26.3103782Z ##[endgroup] 2025-12-04T09:15:26.3105632Z Secret source: Actions 2025-12-04T09:15:26.3106279Z Prepare workflow directory 2025-12-04T09:15:26.3598668Z Prepare all required actions 2025-12-04T09:15:26.3635018Z Getting action download info 2025-12-04T09:15:26.7214596Z Download action repository 'pytorch/test-infra@main' (SHA:39aa74d619174326f4e2fb0e216151c2f29d9ffd) 2025-12-04T09:15:29.1463275Z Download action repository 'pytorch/pytorch@main' (SHA:7716da9fb23f27a65b41f9f016a2afadf281c18f) 2025-12-04T09:15:45.2092454Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-12-04T09:15:45.5539791Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-12-04T09:15:45.8411772Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-12-04T09:15:46.0365291Z Download action repository 'seemethere/download-artifact-s3@1da556a7aa0a088e3153970611f6c432d58e80e6' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:15:46.2720574Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T09:15:46.5703325Z Getting action download info 2025-12-04T09:15:46.6897340Z Download action repository 'actions/checkout@v4' (SHA:34e114876b0b11c390a56381ad16ebd13914f8d5) 2025-12-04T09:15:47.0191622Z Getting action download info 2025-12-04T09:15:47.1477666Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-12-04T09:15:47.3688641Z Getting action download info 2025-12-04T09:15:47.4982134Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-12-04T09:15:47.6997047Z Getting action download info 2025-12-04T09:15:47.9077874Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32) 2025-12-04T09:15:47.9081777Z ##[group] Inputs 2025-12-04T09:15:47.9082183Z build-environment: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck 2025-12-04T09:15:47.9091556Z test-matrix: {"include": [{"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}]} 2025-12-04T09:15:47.9100601Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:15:47.9101400Z sync-tag: 2025-12-04T09:15:47.9102125Z timeout-minutes: 300 2025-12-04T09:15:47.9102383Z use-gha: 2025-12-04T09:15:47.9102599Z dashboard-tag: 2025-12-04T09:15:47.9102837Z s3-bucket: gha-artifacts 2025-12-04T09:15:47.9103106Z aws-role-to-assume: 2025-12-04T09:15:47.9103609Z disable-monitor: false 2025-12-04T09:15:47.9103900Z monitor-log-interval: 5 2025-12-04T09:15:47.9104215Z monitor-data-collect-interval: 1 2025-12-04T09:15:47.9104525Z ##[endgroup] 2025-12-04T09:15:47.9105214Z Complete job name: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:15:47.9787459Z A job started hook has been configured by the self-hosted runner administrator 2025-12-04T09:15:47.9884952Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-12-04T09:15:47.9896386Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:15:47.9896984Z ##[endgroup] 2025-12-04T09:15:49.5003356Z Runner Type: linux.g5.4xlarge.nvidia.gpu 2025-12-04T09:15:49.5003822Z Instance Type: g5.4xlarge 2025-12-04T09:15:49.5004097Z AMI Name: unknown 2025-12-04T09:15:49.5057440Z AMI ID: ami-08982f1c5bf93d976 2025-12-04T09:15:55.0209276Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-12-04T09:15:55.0209690Z with: 2025-12-04T09:15:55.0210196Z github-secret: *** 2025-12-04T09:15:55.0210881Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-12-04T09:15:55.0211619Z activate-with-label: false 2025-12-04T09:15:55.0211897Z label: with-ssh 2025-12-04T09:15:55.0212150Z remove-existing-keys: true 2025-12-04T09:15:55.0212417Z fail-silently: true 2025-12-04T09:15:55.0212654Z env: 2025-12-04T09:15:55.0212864Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:15:55.0213122Z ##[endgroup] 2025-12-04T09:15:55.1612908Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-12-04T09:15:55.1613963Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-12-04T09:15:55.1794620Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-12-04T09:15:55.1795029Z with: 2025-12-04T09:15:55.1795238Z no-sudo: true 2025-12-04T09:15:55.1795472Z submodules: recursive 2025-12-04T09:15:55.1795717Z fetch-depth: 0 2025-12-04T09:15:55.1796145Z env: 2025-12-04T09:15:55.1796353Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:15:55.1796617Z ##[endgroup] 2025-12-04T09:15:55.1871247Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:15:55.1872161Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:15:55.1886829Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:15:55.1887199Z env: 2025-12-04T09:15:55.1887436Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:15:55.1887791Z ##[endgroup] 2025-12-04T09:15:55.1977344Z ##[group]Run # Use all available CPUs for fetching 2025-12-04T09:15:55.1977768Z # Use all available CPUs for fetching 2025-12-04T09:15:55.1978103Z cd "${GITHUB_WORKSPACE}" 2025-12-04T09:15:55.1978438Z git config --global fetch.parallel 0 2025-12-04T09:15:55.1978826Z git config --global submodule.fetchJobs 0 2025-12-04T09:15:55.1979175Z  2025-12-04T09:15:55.1979533Z # Clean workspace. The default checkout action should also do this, but 2025-12-04T09:15:55.1979990Z # do it here as well just in case 2025-12-04T09:15:55.1980308Z if [[ -d .git ]]; then 2025-12-04T09:15:55.1980590Z  if [ -z "${NO_SUDO}" ]; then 2025-12-04T09:15:55.1980896Z  sudo git clean -ffdx 2025-12-04T09:15:55.1981173Z  else 2025-12-04T09:15:55.1981400Z  git clean -ffdx 2025-12-04T09:15:55.1981912Z  fi 2025-12-04T09:15:55.1982208Z fi 2025-12-04T09:15:55.1991567Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:15:55.1992051Z env: 2025-12-04T09:15:55.2011794Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:15:55.2012122Z NO_SUDO: true 2025-12-04T09:15:55.2012386Z ##[endgroup] 2025-12-04T09:15:55.2149040Z ##[group]Run actions/checkout@v4 2025-12-04T09:15:55.2149345Z with: 2025-12-04T09:15:55.2149601Z ref: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:15:55.2149927Z fetch-depth: 0 2025-12-04T09:15:55.2150158Z submodules: recursive 2025-12-04T09:15:55.2150414Z show-progress: false 2025-12-04T09:15:55.2150684Z repository: pytorch/pytorch 2025-12-04T09:15:55.2151079Z token: *** 2025-12-04T09:15:55.2151292Z ssh-strict: true 2025-12-04T09:15:55.2151526Z ssh-user: git 2025-12-04T09:15:55.2151771Z persist-credentials: true 2025-12-04T09:15:55.2152030Z clean: true 2025-12-04T09:15:55.2152290Z sparse-checkout-cone-mode: true 2025-12-04T09:15:55.2152585Z fetch-tags: false 2025-12-04T09:15:55.2152805Z lfs: false 2025-12-04T09:15:55.2153039Z set-safe-directory: true 2025-12-04T09:15:55.2153305Z env: 2025-12-04T09:15:55.2153511Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:15:55.2153764Z ##[endgroup] 2025-12-04T09:15:55.3273582Z Syncing repository: pytorch/pytorch 2025-12-04T09:15:55.3274986Z ##[group]Getting Git version info 2025-12-04T09:15:55.3275488Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T09:15:55.3276139Z [command]/usr/bin/git version 2025-12-04T09:15:55.3491711Z git version 2.50.1 2025-12-04T09:15:55.3538438Z ##[endgroup] 2025-12-04T09:15:55.3548990Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/67c2b306-3cc9-41d5-b516-c83e19055244/.gitconfig' 2025-12-04T09:15:55.3571043Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/67c2b306-3cc9-41d5-b516-c83e19055244' before making global git config changes 2025-12-04T09:15:55.3571963Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T09:15:55.3576781Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:15:55.3629987Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T09:15:55.3641706Z ##[group]Initializing the repository 2025-12-04T09:15:55.3645925Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:15:55.3724716Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-12-04T09:15:55.3725325Z hint: is subject to change. To configure the initial branch name to use in all 2025-12-04T09:15:55.3725898Z hint: of your new repositories, which will suppress this warning, call: 2025-12-04T09:15:55.3726326Z hint: 2025-12-04T09:15:55.3726623Z hint: git config --global init.defaultBranch 2025-12-04T09:15:55.3726981Z hint: 2025-12-04T09:15:55.3727320Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-12-04T09:15:55.3727959Z hint: 'development'. The just-created branch can be renamed via this command: 2025-12-04T09:15:55.3728383Z hint: 2025-12-04T09:15:55.3728611Z hint: git branch -m 2025-12-04T09:15:55.3728863Z hint: 2025-12-04T09:15:55.3729227Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-12-04T09:15:55.3736575Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-12-04T09:15:55.3748545Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-12-04T09:15:55.3798526Z ##[endgroup] 2025-12-04T09:15:55.3799087Z ##[group]Disabling automatic garbage collection 2025-12-04T09:15:55.3802185Z [command]/usr/bin/git config --local gc.auto 0 2025-12-04T09:15:55.3836546Z ##[endgroup] 2025-12-04T09:15:55.3837058Z ##[group]Setting up auth 2025-12-04T09:15:55.3843311Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T09:15:55.3877743Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T09:15:55.4321666Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T09:15:55.4356929Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T09:15:55.4758345Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:15:55.4795834Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T09:15:55.5179615Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T09:15:55.5232024Z ##[endgroup] 2025-12-04T09:15:55.5232465Z ##[group]Fetching the repository 2025-12-04T09:15:55.5240101Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-12-04T09:16:46.9838176Z From https://github.com/pytorch/pytorch 2025-12-04T09:16:46.9838957Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-12-04T09:16:46.9839639Z * [new branch] 2.9.1 -> origin/2.9.1 2025-12-04T09:16:46.9840297Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-12-04T09:16:46.9841387Z * [new branch] Flamefire-patch-1 -> origin/Flamefire-patch-1 2025-12-04T09:16:46.9843228Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-12-04T09:16:46.9845018Z * [new branch] HOPrintFunc -> origin/HOPrintFunc 2025-12-04T09:16:46.9848349Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-12-04T09:16:46.9851227Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-12-04T09:16:46.9852939Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-12-04T09:16:46.9854969Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-12-04T09:16:46.9856783Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-12-04T09:16:46.9858607Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-12-04T09:16:46.9860457Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-12-04T09:16:46.9862444Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-12-04T09:16:46.9864906Z * [new branch] VLA_exp -> origin/VLA_exp 2025-12-04T09:16:46.9867030Z * [new branch] activation_bench -> origin/activation_bench 2025-12-04T09:16:46.9869665Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-12-04T09:16:46.9872131Z * [new branch] adi/onednn_aarch64 -> origin/adi/onednn_aarch64 2025-12-04T09:16:46.9873445Z * [new branch] adi/test -> origin/adi/test 2025-12-04T09:16:46.9875708Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-12-04T09:16:46.9877424Z * [new branch] adi/test_m8g -> origin/adi/test_m8g 2025-12-04T09:16:46.9879412Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-12-04T09:16:46.9881039Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-12-04T09:16:46.9883088Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-12-04T09:16:46.9884540Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-12-04T09:16:46.9887059Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-12-04T09:16:46.9890414Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-12-04T09:16:46.9892048Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-12-04T09:16:46.9894613Z * [new branch] albanD-patch-1 -> origin/albanD-patch-1 2025-12-04T09:16:46.9895989Z * [new branch] also-surround-shimh -> origin/also-surround-shimh 2025-12-04T09:16:46.9899009Z * [new branch] angelayi/aot_compile -> origin/angelayi/aot_compile 2025-12-04T09:16:46.9900741Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-12-04T09:16:46.9902930Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-12-04T09:16:46.9904837Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-12-04T09:16:46.9906367Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-12-04T09:16:46.9908643Z * [new branch] angelayi/inductor_const -> origin/angelayi/inductor_const 2025-12-04T09:16:46.9910199Z * [new branch] angelayi/lstm -> origin/angelayi/lstm 2025-12-04T09:16:46.9912954Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-12-04T09:16:46.9915426Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-12-04T09:16:46.9917401Z * [new branch] angelayi/side_eff -> origin/angelayi/side_eff 2025-12-04T09:16:46.9919229Z * [new branch] angelayi/state_dict -> origin/angelayi/state_dict 2025-12-04T09:16:46.9921366Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-12-04T09:16:46.9923341Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-12-04T09:16:46.9925129Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-12-04T09:16:46.9927272Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-12-04T09:16:46.9928943Z * [new branch] annotate_assert -> origin/annotate_assert 2025-12-04T09:16:46.9931100Z * [new branch] annotate_fallback_kernel -> origin/annotate_fallback_kernel 2025-12-04T09:16:46.9932749Z * [new branch] annotation_deepcopy -> origin/annotation_deepcopy 2025-12-04T09:16:46.9934990Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-12-04T09:16:46.9937109Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-12-04T09:16:46.9939280Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-12-04T09:16:46.9941175Z * [new branch] aoti_const_device -> origin/aoti_const_device 2025-12-04T09:16:46.9943022Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-12-04T09:16:46.9944622Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-12-04T09:16:46.9946722Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-12-04T09:16:46.9950088Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-12-04T09:16:46.9951581Z * [new branch] async_tp -> origin/async_tp 2025-12-04T09:16:46.9953933Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-12-04T09:16:46.9955618Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-12-04T09:16:46.9957902Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-12-04T09:16:46.9959986Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-12-04T09:16:46.9961548Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-12-04T09:16:46.9963874Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-12-04T09:16:46.9965763Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-12-04T09:16:46.9967746Z * [new branch] atalman-patch-7 -> origin/atalman-patch-7 2025-12-04T09:16:46.9969862Z * [new branch] atalman-patch-8 -> origin/atalman-patch-8 2025-12-04T09:16:46.9971447Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-12-04T09:16:46.9973580Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-12-04T09:16:46.9975562Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-12-04T09:16:46.9977648Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-12-04T09:16:46.9980103Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-12-04T09:16:46.9981536Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-12-04T09:16:46.9983664Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-12-04T09:16:46.9985228Z * [new branch] bahuang/test -> origin/bahuang/test 2025-12-04T09:16:46.9988197Z * [new branch] base/1.5 -> origin/base/1.5 2025-12-04T09:16:46.9990375Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-12-04T09:16:46.9992018Z * [new branch] bench_scaled_mm_ops -> origin/bench_scaled_mm_ops 2025-12-04T09:16:46.9994389Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-12-04T09:16:46.9995843Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-12-04T09:16:46.9998625Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-12-04T09:16:47.0001242Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-12-04T09:16:47.0004098Z * [new branch] bf/bug-static-input -> origin/bf/bug-static-input 2025-12-04T09:16:47.0005492Z * [new branch] bf/cg-backend -> origin/bf/cg-backend 2025-12-04T09:16:47.0007637Z * [new branch] bf/cg-nccl-test -> origin/bf/cg-nccl-test 2025-12-04T09:16:47.0009152Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-12-04T09:16:47.0011371Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-12-04T09:16:47.0012836Z * [new branch] bf/combo-debug-log -> origin/bf/combo-debug-log 2025-12-04T09:16:47.0048152Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-12-04T09:16:47.0049171Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-12-04T09:16:47.0050247Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-12-04T09:16:47.0051289Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-12-04T09:16:47.0051969Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-12-04T09:16:47.0052752Z * [new branch] bf/dynamo-partition -> origin/bf/dynamo-partition 2025-12-04T09:16:47.0053418Z * [new branch] bf/lite -> origin/bf/lite 2025-12-04T09:16:47.0053931Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-12-04T09:16:47.0054554Z * [new branch] bf/partition-cache-free-symbols -> origin/bf/partition-cache-free-symbols 2025-12-04T09:16:47.0055286Z * [new branch] bf/partition-memory-plan -> origin/bf/partition-memory-plan 2025-12-04T09:16:47.0055888Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-12-04T09:16:47.0056473Z * [new branch] bf/partition-view-fallback -> origin/bf/partition-view-fallback 2025-12-04T09:16:47.0057069Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-12-04T09:16:47.0057707Z * [new branch] bf/timm-nov-26-2025 -> origin/bf/timm-nov-26-2025 2025-12-04T09:16:47.0058270Z * [new branch] bf/transformer-pin-4-57-3 -> origin/bf/transformer-pin-4-57-3 2025-12-04T09:16:47.0058864Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-12-04T09:16:47.0059502Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-12-04T09:16:47.0060127Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-12-04T09:16:47.0060698Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-12-04T09:16:47.0061281Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-12-04T09:16:47.0061920Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-12-04T09:16:47.0062714Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-12-04T09:16:47.0063306Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-12-04T09:16:47.0063892Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-12-04T09:16:47.0064834Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-12-04T09:16:47.0066768Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-12-04T09:16:47.0068883Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-12-04T09:16:47.0071240Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-12-04T09:16:47.0073445Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-12-04T09:16:47.0075707Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-12-04T09:16:47.0078070Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-12-04T09:16:47.0081188Z * [new branch] brister/fx_device_type -> origin/brister/fx_device_type 2025-12-04T09:16:47.0083414Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-12-04T09:16:47.0085721Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-12-04T09:16:47.0088119Z * [new branch] bwd-backup -> origin/bwd-backup 2025-12-04T09:16:47.0090620Z * [new branch] c57382a49 -> origin/c57382a49 2025-12-04T09:16:47.0092886Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-12-04T09:16:47.0095150Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-12-04T09:16:47.0098313Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-12-04T09:16:47.0100691Z * [new branch] cccclai-patch-1 -> origin/cccclai-patch-1 2025-12-04T09:16:47.0103193Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0105530Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0107961Z * [new branch] cherry-pick-162208-by-pytorch_bot_bot_ -> origin/cherry-pick-162208-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0110365Z * [new branch] cherry-pick-163169-by-pytorch_bot_bot_ -> origin/cherry-pick-163169-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0112748Z * [new branch] cherry-pick-165086-by-pytorch_bot_bot_ -> origin/cherry-pick-165086-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0115240Z * [new branch] cherry-pick-165514-by-pytorch_bot_bot_ -> origin/cherry-pick-165514-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0117592Z * [new branch] cherry-pick-165601-by-pytorch_bot_bot_ -> origin/cherry-pick-165601-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0120005Z * [new branch] cherry-pick-165667-by-pytorch_bot_bot_ -> origin/cherry-pick-165667-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0122461Z * [new branch] cherry-pick-165815-by-pytorch_bot_bot_ -> origin/cherry-pick-165815-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0124873Z * [new branch] cherry-pick-165922-by-pytorch_bot_bot_ -> origin/cherry-pick-165922-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0127287Z * [new branch] cherry-pick-166148-by-pytorch_bot_bot_ -> origin/cherry-pick-166148-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0129770Z * [new branch] cherry-pick-166181-by-pytorch_bot_bot_ -> origin/cherry-pick-166181-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0132117Z * [new branch] cherry-pick-166404-by-pytorch_bot_bot_ -> origin/cherry-pick-166404-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0134507Z * [new branch] cherry-pick-166427-by-pytorch_bot_bot_ -> origin/cherry-pick-166427-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0137384Z * [new branch] cherry-pick-166480-by-pytorch_bot_bot_ -> origin/cherry-pick-166480-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0139658Z * [new branch] cherry-pick-166570-by-pytorch_bot_bot_ -> origin/cherry-pick-166570-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0141960Z * [new branch] cherry-pick-166993-by-pytorch_bot_bot_ -> origin/cherry-pick-166993-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0144473Z * [new branch] cherry-pick-167111-by-pytorch_bot_bot_ -> origin/cherry-pick-167111-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0146897Z * [new branch] cherry-pick-167478-by-pytorch_bot_bot_ -> origin/cherry-pick-167478-by-pytorch_bot_bot_ 2025-12-04T09:16:47.0149088Z * [new branch] cherry_pick_166036_166040 -> origin/cherry_pick_166036_166040 2025-12-04T09:16:47.0151465Z * [new branch] cherry_pick_166457 -> origin/cherry_pick_166457 2025-12-04T09:16:47.0153907Z * [new branch] cherrypick_166338 -> origin/cherrypick_166338 2025-12-04T09:16:47.0156319Z * [new branch] cherrypick_166458 -> origin/cherrypick_166458 2025-12-04T09:16:47.0158580Z * [new branch] cherrypick_166586 -> origin/cherrypick_166586 2025-12-04T09:16:47.0160918Z * [new branch] cherrypick_166956 -> origin/cherrypick_166956 2025-12-04T09:16:47.0163284Z * [new branch] ci_attn -> origin/ci_attn 2025-12-04T09:16:47.0165614Z * [new branch] codex-testing -> origin/codex-testing 2025-12-04T09:16:47.0169103Z * [new branch] codex/add-check_memory_overlap-helper-functions -> origin/codex/add-check_memory_overlap-helper-functions 2025-12-04T09:16:47.0171197Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-12-04T09:16:47.0174199Z * [new branch] codex/investigate-segfaults-in-get_tensor_storage_id -> origin/codex/investigate-segfaults-in-get_tensor_storage_id 2025-12-04T09:16:47.0176763Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-12-04T09:16:47.0178848Z * [new branch] compatiblpy39util -> origin/compatiblpy39util 2025-12-04T09:16:47.0181206Z * [new branch] cond_hop_device -> origin/cond_hop_device 2025-12-04T09:16:47.0183566Z * [new branch] context_test -> origin/context_test 2025-12-04T09:16:47.0186763Z * [new branch] copilot/code-style-cleanup-python-pip -> origin/copilot/code-style-cleanup-python-pip 2025-12-04T09:16:47.0189655Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-12-04T09:16:47.0192282Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-12-04T09:16:47.0195468Z * [new branch] crpa/typo-in-inductor_comm_lowering -> origin/crpa/typo-in-inductor_comm_lowering 2025-12-04T09:16:47.0198364Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-12-04T09:16:47.0200626Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-12-04T09:16:47.0203329Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-12-04T09:16:47.0205642Z * [new branch] csl/clean_up -> origin/csl/clean_up 2025-12-04T09:16:47.0208012Z * [new branch] csl/fix_retry_segfault_exit -> origin/csl/fix_retry_segfault_exit 2025-12-04T09:16:47.0210229Z * [new branch] csl/katex -> origin/csl/katex 2025-12-04T09:16:47.0212799Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-12-04T09:16:47.0215475Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-12-04T09:16:47.0218135Z * [new branch] csl/lint_thing -> origin/csl/lint_thing 2025-12-04T09:16:47.0220589Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-12-04T09:16:47.0222984Z * [new branch] csl/manually_gen_json -> origin/csl/manually_gen_json 2025-12-04T09:16:47.0225266Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-12-04T09:16:47.0227686Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-12-04T09:16:47.0230115Z * [new branch] csl/print_timing -> origin/csl/print_timing 2025-12-04T09:16:47.0232445Z * [new branch] csl/remove_experiment -> origin/csl/remove_experiment 2025-12-04T09:16:47.0234850Z * [new branch] csl/remove_maybe_unused_var -> origin/csl/remove_maybe_unused_var 2025-12-04T09:16:47.0237692Z * [new branch] csl/remove_repo_specific_autolabel -> origin/csl/remove_repo_specific_autolabel 2025-12-04T09:16:47.0240040Z * [new branch] csl/remove_run_parallel -> origin/csl/remove_run_parallel 2025-12-04T09:16:47.0242209Z * [new branch] csl/remove_unused_vars -> origin/csl/remove_unused_vars 2025-12-04T09:16:47.0244675Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-12-04T09:16:47.0246948Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-12-04T09:16:47.0249445Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-12-04T09:16:47.0251707Z * [new branch] csl/td_job_level -> origin/csl/td_job_level 2025-12-04T09:16:47.0254096Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-12-04T09:16:47.0256587Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-12-04T09:16:47.0258828Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-12-04T09:16:47.0261130Z * [new branch] csl/upload_json_running -> origin/csl/upload_json_running 2025-12-04T09:16:47.0263499Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-12-04T09:16:47.0265783Z * [new branch] csl/xml_stuff -> origin/csl/xml_stuff 2025-12-04T09:16:47.0268163Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-12-04T09:16:47.0270552Z * [new branch] cuda_mempool -> origin/cuda_mempool 2025-12-04T09:16:47.0272892Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-12-04T09:16:47.0275896Z * [new branch] d4l3k/debug_plane_frtrace -> origin/d4l3k/debug_plane_frtrace 2025-12-04T09:16:47.0278819Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-12-04T09:16:47.0281139Z * [new branch] debug-guard -> origin/debug-guard 2025-12-04T09:16:47.0283608Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-12-04T09:16:47.0290291Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-12-04T09:16:47.0293201Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 2025-12-04T09:16:47.0295922Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-12-04T09:16:47.0298171Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-12-04T09:16:47.0301675Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-12-04T09:16:47.0305082Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-12-04T09:16:47.0307752Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-12-04T09:16:47.0310233Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-12-04T09:16:47.0312525Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-12-04T09:16:47.0314933Z * [new branch] dev/joona/fix_sdpa_memtest -> origin/dev/joona/fix_sdpa_memtest 2025-12-04T09:16:47.0317492Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-12-04T09:16:47.0320066Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-12-04T09:16:47.0323014Z * [new branch] dev/joona/scalar_clamp -> origin/dev/joona/scalar_clamp 2025-12-04T09:16:47.0325877Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-12-04T09:16:47.0329059Z * [new branch] dev/joona/sdpa_api -> origin/dev/joona/sdpa_api 2025-12-04T09:16:47.0331590Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-12-04T09:16:47.0334223Z * [new branch] dev/joona/ulpAssertClose -> origin/dev/joona/ulpAssertClose 2025-12-04T09:16:47.0337160Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-12-04T09:16:47.0339305Z * [new branch] disp_counter -> origin/disp_counter 2025-12-04T09:16:47.0341687Z * [new branch] divyanshk-patch-1 -> origin/divyanshk-patch-1 2025-12-04T09:16:47.0343895Z * [new branch] docs -> origin/docs 2025-12-04T09:16:47.0346353Z * [new branch] documentation -> origin/documentation 2025-12-04T09:16:47.0348680Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-12-04T09:16:47.0351772Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-12-04T09:16:47.0353958Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-12-04T09:16:47.0356090Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-12-04T09:16:47.0358419Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-12-04T09:16:47.0360821Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-12-04T09:16:47.0363231Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-12-04T09:16:47.0365603Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-12-04T09:16:47.0368077Z * [new branch] eqy-patch-5 -> origin/eqy-patch-5 2025-12-04T09:16:47.0370524Z * [new branch] eqy-patch-6 -> origin/eqy-patch-6 2025-12-04T09:16:47.0373502Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-12-04T09:16:47.0375906Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-12-04T09:16:47.0378075Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-12-04T09:16:47.0380343Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-12-04T09:16:47.0382820Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-12-04T09:16:47.0385431Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-12-04T09:16:47.0388089Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-12-04T09:16:47.0390313Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-12-04T09:16:47.0392946Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-12-04T09:16:47.0395066Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-12-04T09:16:47.0397348Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-12-04T09:16:47.0399891Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-12-04T09:16:47.0402016Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-12-04T09:16:47.0404384Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-12-04T09:16:47.0406915Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-12-04T09:16:47.0409453Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-12-04T09:16:47.0411743Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-12-04T09:16:47.0414084Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-12-04T09:16:47.0416516Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-12-04T09:16:47.0418931Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-12-04T09:16:47.0421097Z * [new branch] exec -> origin/exec 2025-12-04T09:16:47.0423633Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-12-04T09:16:47.0426067Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-12-04T09:16:47.0428421Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-12-04T09:16:47.0430860Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-12-04T09:16:47.0433177Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-12-04T09:16:47.0435688Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-12-04T09:16:47.0438394Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-12-04T09:16:47.0440777Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-12-04T09:16:47.0443112Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-12-04T09:16:47.0445372Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-12-04T09:16:47.0447722Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-12-04T09:16:47.0450216Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-12-04T09:16:47.0452563Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-12-04T09:16:47.0454882Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-12-04T09:16:47.0457298Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-12-04T09:16:47.0459640Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-12-04T09:16:47.0461970Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-12-04T09:16:47.0464315Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-12-04T09:16:47.0466644Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-12-04T09:16:47.0468903Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-12-04T09:16:47.0471821Z * [new branch] export-D84373821 -> origin/export-D84373821 2025-12-04T09:16:47.0474374Z * [new branch] export-D84612194 -> origin/export-D84612194 2025-12-04T09:16:47.0476558Z * [new branch] export-D84890985 -> origin/export-D84890985 2025-12-04T09:16:47.0478830Z * [new branch] export-D85122326 -> origin/export-D85122326 2025-12-04T09:16:47.0481301Z * [new branch] export-D86256198 -> origin/export-D86256198 2025-12-04T09:16:47.0483605Z * [new branch] export-D86460608 -> origin/export-D86460608 2025-12-04T09:16:47.0486037Z * [new branch] export-D86474796 -> origin/export-D86474796 2025-12-04T09:16:47.0488616Z * [new branch] export-D86712396 -> origin/export-D86712396 2025-12-04T09:16:47.0491046Z * [new branch] export-D87022129 -> origin/export-D87022129 2025-12-04T09:16:47.0493408Z * [new branch] export-D87838959 -> origin/export-D87838959 2025-12-04T09:16:47.0496358Z * [new branch] export-D88319437 -> origin/export-D88319437 2025-12-04T09:16:47.0498913Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-12-04T09:16:47.0501209Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-12-04T09:16:47.0503533Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-12-04T09:16:47.0505768Z * [new branch] ezyang-war -> origin/ezyang-war 2025-12-04T09:16:47.0508848Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-12-04T09:16:47.0511068Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-12-04T09:16:47.0514200Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-12-04T09:16:47.0516480Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-12-04T09:16:47.0519618Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-12-04T09:16:47.0521988Z * [new branch] fca -> origin/fca 2025-12-04T09:16:47.0524334Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-12-04T09:16:47.0526716Z * [new branch] fca5 -> origin/fca5 2025-12-04T09:16:47.0529848Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-12-04T09:16:47.0532085Z * [new branch] feature/numa-forkserver -> origin/feature/numa-forkserver 2025-12-04T09:16:47.0534928Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-12-04T09:16:47.0538481Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-12-04T09:16:47.0540916Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-12-04T09:16:47.0543155Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-12-04T09:16:47.0545408Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-12-04T09:16:47.0547419Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-12-04T09:16:47.0549325Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-12-04T09:16:47.0551886Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-12-04T09:16:47.0554183Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-12-04T09:16:47.0556181Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-12-04T09:16:47.0558556Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-12-04T09:16:47.0561224Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-12-04T09:16:47.0563291Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-12-04T09:16:47.0565653Z * [new branch] fix_addmm_issue -> origin/fix_addmm_issue 2025-12-04T09:16:47.0568065Z * [new branch] fix_amd_missing_cluster_dims -> origin/fix_amd_missing_cluster_dims 2025-12-04T09:16:47.0570469Z * [new branch] fix_bench_bwd_pass -> origin/fix_bench_bwd_pass 2025-12-04T09:16:47.0572727Z * [new branch] fix_mem_profiler_config -> origin/fix_mem_profiler_config 2025-12-04T09:16:47.0575007Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-12-04T09:16:47.0577297Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-12-04T09:16:47.0580252Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-12-04T09:16:47.0582639Z * [new branch] fixes-triage -> origin/fixes-triage 2025-12-04T09:16:47.0585341Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-12-04T09:16:47.0587654Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-12-04T09:16:47.0589944Z * [new branch] flex-flash -> origin/flex-flash 2025-12-04T09:16:47.0592324Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-12-04T09:16:47.0594727Z * [new branch] flex_flash -> origin/flex_flash 2025-12-04T09:16:47.0597828Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-12-04T09:16:47.0600142Z * [new branch] fmassa/tests_comm_compute_scheduler -> origin/fmassa/tests_comm_compute_scheduler 2025-12-04T09:16:47.0602358Z * [new branch] forkserver_fix -> origin/forkserver_fix 2025-12-04T09:16:47.0604666Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-12-04T09:16:47.0607075Z * [new branch] fx_cpp -> origin/fx_cpp 2025-12-04T09:16:47.0610329Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-12-04T09:16:47.0612730Z * [new branch] galv-patch-1 -> origin/galv-patch-1 2025-12-04T09:16:47.0616017Z * [new branch] galv/cudagraphs-conditional-nodes-4 -> origin/galv/cudagraphs-conditional-nodes-4 2025-12-04T09:16:47.0619133Z * [new branch] georgehong/cmakelists-patch -> origin/georgehong/cmakelists-patch 2025-12-04T09:16:47.0623477Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-12-04T09:16:47.0625822Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-12-04T09:16:47.0629479Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-12-04T09:16:47.0631681Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-12-04T09:16:47.0635842Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-12-04T09:16:47.0638280Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-12-04T09:16:47.0641956Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-12-04T09:16:47.0644117Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-12-04T09:16:47.0646396Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-12-04T09:16:47.0649808Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-12-04T09:16:47.0651949Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-12-04T09:16:47.0654103Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-12-04T09:16:47.0657346Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-12-04T09:16:47.0659437Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-12-04T09:16:47.0661710Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-12-04T09:16:47.0664593Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-12-04T09:16:47.0666873Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-12-04T09:16:47.0669149Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-12-04T09:16:47.0672414Z * [new branch] gh/H-Huang/226/base -> origin/gh/H-Huang/226/base 2025-12-04T09:16:47.0674757Z * [new branch] gh/H-Huang/226/head -> origin/gh/H-Huang/226/head 2025-12-04T09:16:47.0676922Z * [new branch] gh/H-Huang/226/orig -> origin/gh/H-Huang/226/orig 2025-12-04T09:16:47.0679972Z * [new branch] gh/H-Huang/228/base -> origin/gh/H-Huang/228/base 2025-12-04T09:16:47.0682482Z * [new branch] gh/H-Huang/228/head -> origin/gh/H-Huang/228/head 2025-12-04T09:16:47.0684527Z * [new branch] gh/H-Huang/228/orig -> origin/gh/H-Huang/228/orig 2025-12-04T09:16:47.0688184Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-12-04T09:16:47.0690489Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-12-04T09:16:47.0692636Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-12-04T09:16:47.0695769Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-12-04T09:16:47.0698213Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-12-04T09:16:47.0700561Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-12-04T09:16:47.0703562Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-12-04T09:16:47.0705898Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-12-04T09:16:47.0708191Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-12-04T09:16:47.0711252Z * [new branch] gh/IvanKobzarev/162/base -> origin/gh/IvanKobzarev/162/base 2025-12-04T09:16:47.0713788Z * [new branch] gh/IvanKobzarev/162/head -> origin/gh/IvanKobzarev/162/head 2025-12-04T09:16:47.0715985Z * [new branch] gh/IvanKobzarev/162/orig -> origin/gh/IvanKobzarev/162/orig 2025-12-04T09:16:47.0719191Z * [new branch] gh/IvanKobzarev/163/base -> origin/gh/IvanKobzarev/163/base 2025-12-04T09:16:47.0721342Z * [new branch] gh/IvanKobzarev/163/head -> origin/gh/IvanKobzarev/163/head 2025-12-04T09:16:47.0723609Z * [new branch] gh/IvanKobzarev/163/orig -> origin/gh/IvanKobzarev/163/orig 2025-12-04T09:16:47.0726710Z * [new branch] gh/IvanKobzarev/166/base -> origin/gh/IvanKobzarev/166/base 2025-12-04T09:16:47.0729155Z * [new branch] gh/IvanKobzarev/166/head -> origin/gh/IvanKobzarev/166/head 2025-12-04T09:16:47.0731413Z * [new branch] gh/IvanKobzarev/166/orig -> origin/gh/IvanKobzarev/166/orig 2025-12-04T09:16:47.0734471Z * [new branch] gh/IvanKobzarev/167/base -> origin/gh/IvanKobzarev/167/base 2025-12-04T09:16:47.0736960Z * [new branch] gh/IvanKobzarev/167/head -> origin/gh/IvanKobzarev/167/head 2025-12-04T09:16:47.0739233Z * [new branch] gh/IvanKobzarev/167/orig -> origin/gh/IvanKobzarev/167/orig 2025-12-04T09:16:47.0742159Z * [new branch] gh/IvanKobzarev/168/base -> origin/gh/IvanKobzarev/168/base 2025-12-04T09:16:47.0744617Z * [new branch] gh/IvanKobzarev/168/head -> origin/gh/IvanKobzarev/168/head 2025-12-04T09:16:47.0746764Z * [new branch] gh/IvanKobzarev/168/orig -> origin/gh/IvanKobzarev/168/orig 2025-12-04T09:16:47.0749799Z * [new branch] gh/IvanKobzarev/169/base -> origin/gh/IvanKobzarev/169/base 2025-12-04T09:16:47.0752095Z * [new branch] gh/IvanKobzarev/169/head -> origin/gh/IvanKobzarev/169/head 2025-12-04T09:16:47.0754343Z * [new branch] gh/IvanKobzarev/169/orig -> origin/gh/IvanKobzarev/169/orig 2025-12-04T09:16:47.0757261Z * [new branch] gh/IvanKobzarev/170/base -> origin/gh/IvanKobzarev/170/base 2025-12-04T09:16:47.0759502Z * [new branch] gh/IvanKobzarev/170/head -> origin/gh/IvanKobzarev/170/head 2025-12-04T09:16:47.0761741Z * [new branch] gh/IvanKobzarev/170/orig -> origin/gh/IvanKobzarev/170/orig 2025-12-04T09:16:47.0764982Z * [new branch] gh/IvanKobzarev/171/base -> origin/gh/IvanKobzarev/171/base 2025-12-04T09:16:47.0767247Z * [new branch] gh/IvanKobzarev/171/head -> origin/gh/IvanKobzarev/171/head 2025-12-04T09:16:47.0769687Z * [new branch] gh/IvanKobzarev/171/orig -> origin/gh/IvanKobzarev/171/orig 2025-12-04T09:16:47.0772686Z * [new branch] gh/IvanKobzarev/172/base -> origin/gh/IvanKobzarev/172/base 2025-12-04T09:16:47.0775226Z * [new branch] gh/IvanKobzarev/172/head -> origin/gh/IvanKobzarev/172/head 2025-12-04T09:16:47.0777449Z * [new branch] gh/IvanKobzarev/172/orig -> origin/gh/IvanKobzarev/172/orig 2025-12-04T09:16:47.0780479Z * [new branch] gh/IvanKobzarev/173/base -> origin/gh/IvanKobzarev/173/base 2025-12-04T09:16:47.0782785Z * [new branch] gh/IvanKobzarev/173/head -> origin/gh/IvanKobzarev/173/head 2025-12-04T09:16:47.0785057Z * [new branch] gh/IvanKobzarev/173/orig -> origin/gh/IvanKobzarev/173/orig 2025-12-04T09:16:47.0788061Z * [new branch] gh/IvanKobzarev/174/base -> origin/gh/IvanKobzarev/174/base 2025-12-04T09:16:47.0790360Z * [new branch] gh/IvanKobzarev/174/head -> origin/gh/IvanKobzarev/174/head 2025-12-04T09:16:47.0792645Z * [new branch] gh/IvanKobzarev/174/orig -> origin/gh/IvanKobzarev/174/orig 2025-12-04T09:16:47.0795730Z * [new branch] gh/IvanKobzarev/175/base -> origin/gh/IvanKobzarev/175/base 2025-12-04T09:16:47.0798065Z * [new branch] gh/IvanKobzarev/175/head -> origin/gh/IvanKobzarev/175/head 2025-12-04T09:16:47.0800450Z * [new branch] gh/IvanKobzarev/175/orig -> origin/gh/IvanKobzarev/175/orig 2025-12-04T09:16:47.0803641Z * [new branch] gh/IvanKobzarev/176/base -> origin/gh/IvanKobzarev/176/base 2025-12-04T09:16:47.0805928Z * [new branch] gh/IvanKobzarev/176/head -> origin/gh/IvanKobzarev/176/head 2025-12-04T09:16:47.0808287Z * [new branch] gh/IvanKobzarev/176/orig -> origin/gh/IvanKobzarev/176/orig 2025-12-04T09:16:47.0811685Z * [new branch] gh/IvanKobzarev/177/base -> origin/gh/IvanKobzarev/177/base 2025-12-04T09:16:47.0814055Z * [new branch] gh/IvanKobzarev/177/head -> origin/gh/IvanKobzarev/177/head 2025-12-04T09:16:47.0816411Z * [new branch] gh/IvanKobzarev/177/orig -> origin/gh/IvanKobzarev/177/orig 2025-12-04T09:16:47.0819410Z * [new branch] gh/IvanKobzarev/178/base -> origin/gh/IvanKobzarev/178/base 2025-12-04T09:16:47.0821719Z * [new branch] gh/IvanKobzarev/178/head -> origin/gh/IvanKobzarev/178/head 2025-12-04T09:16:47.0824023Z * [new branch] gh/IvanKobzarev/178/orig -> origin/gh/IvanKobzarev/178/orig 2025-12-04T09:16:47.0827129Z * [new branch] gh/IvanKobzarev/179/base -> origin/gh/IvanKobzarev/179/base 2025-12-04T09:16:47.0829333Z * [new branch] gh/IvanKobzarev/179/head -> origin/gh/IvanKobzarev/179/head 2025-12-04T09:16:47.0831824Z * [new branch] gh/IvanKobzarev/179/orig -> origin/gh/IvanKobzarev/179/orig 2025-12-04T09:16:47.0834802Z * [new branch] gh/IvanKobzarev/180/base -> origin/gh/IvanKobzarev/180/base 2025-12-04T09:16:47.0837423Z * [new branch] gh/IvanKobzarev/180/head -> origin/gh/IvanKobzarev/180/head 2025-12-04T09:16:47.0839772Z * [new branch] gh/IvanKobzarev/180/orig -> origin/gh/IvanKobzarev/180/orig 2025-12-04T09:16:47.0843052Z * [new branch] gh/IvanKobzarev/181/base -> origin/gh/IvanKobzarev/181/base 2025-12-04T09:16:47.0845383Z * [new branch] gh/IvanKobzarev/181/head -> origin/gh/IvanKobzarev/181/head 2025-12-04T09:16:47.0847695Z * [new branch] gh/IvanKobzarev/181/orig -> origin/gh/IvanKobzarev/181/orig 2025-12-04T09:16:47.0851033Z * [new branch] gh/IvanKobzarev/182/base -> origin/gh/IvanKobzarev/182/base 2025-12-04T09:16:47.0853310Z * [new branch] gh/IvanKobzarev/182/head -> origin/gh/IvanKobzarev/182/head 2025-12-04T09:16:47.0855663Z * [new branch] gh/IvanKobzarev/182/orig -> origin/gh/IvanKobzarev/182/orig 2025-12-04T09:16:47.0858859Z * [new branch] gh/IvanKobzarev/183/base -> origin/gh/IvanKobzarev/183/base 2025-12-04T09:16:47.0861249Z * [new branch] gh/IvanKobzarev/183/head -> origin/gh/IvanKobzarev/183/head 2025-12-04T09:16:47.0863519Z * [new branch] gh/IvanKobzarev/183/orig -> origin/gh/IvanKobzarev/183/orig 2025-12-04T09:16:47.0866643Z * [new branch] gh/IvanKobzarev/184/base -> origin/gh/IvanKobzarev/184/base 2025-12-04T09:16:47.0869038Z * [new branch] gh/IvanKobzarev/184/head -> origin/gh/IvanKobzarev/184/head 2025-12-04T09:16:47.0871350Z * [new branch] gh/IvanKobzarev/184/orig -> origin/gh/IvanKobzarev/184/orig 2025-12-04T09:16:47.0874914Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-12-04T09:16:47.0877286Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-12-04T09:16:47.0880112Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-12-04T09:16:47.0882394Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-12-04T09:16:47.0885564Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-12-04T09:16:47.0887970Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-12-04T09:16:47.0891104Z * [new branch] gh/NikhilAPatel/5/base -> origin/gh/NikhilAPatel/5/base 2025-12-04T09:16:47.0893338Z * [new branch] gh/NikhilAPatel/5/head -> origin/gh/NikhilAPatel/5/head 2025-12-04T09:16:47.0895654Z * [new branch] gh/NikhilAPatel/5/orig -> origin/gh/NikhilAPatel/5/orig 2025-12-04T09:16:47.0899169Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-12-04T09:16:47.0901436Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-12-04T09:16:47.0903734Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-12-04T09:16:47.0906726Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-12-04T09:16:47.0909056Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-12-04T09:16:47.0911394Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-12-04T09:16:47.0914378Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-12-04T09:16:47.0916663Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-12-04T09:16:47.0919006Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-12-04T09:16:47.0930839Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-12-04T09:16:47.0931630Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-12-04T09:16:47.0932136Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-12-04T09:16:47.0932635Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-12-04T09:16:47.0933326Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-12-04T09:16:47.0934092Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-12-04T09:16:47.0938781Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-12-04T09:16:47.0940987Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-12-04T09:16:47.0943273Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-12-04T09:16:47.0946194Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-12-04T09:16:47.0948497Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-12-04T09:16:47.0950829Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-12-04T09:16:47.0953752Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-12-04T09:16:47.0955884Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-12-04T09:16:47.0958289Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-12-04T09:16:47.0961139Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-12-04T09:16:47.0964131Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-12-04T09:16:47.0966264Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-12-04T09:16:47.0968720Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-12-04T09:16:47.0971723Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-12-04T09:16:47.0973829Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-12-04T09:16:47.0976143Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-12-04T09:16:47.0979129Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-12-04T09:16:47.0981228Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-12-04T09:16:47.0983487Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-12-04T09:16:47.0986858Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-12-04T09:16:47.0989030Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-12-04T09:16:47.0991445Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-12-04T09:16:47.0994927Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-12-04T09:16:47.0997326Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-12-04T09:16:47.0999639Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-12-04T09:16:47.1002624Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-12-04T09:16:47.1004999Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-12-04T09:16:47.1007326Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-12-04T09:16:47.1010711Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-12-04T09:16:47.1014390Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-12-04T09:16:47.1015552Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-12-04T09:16:47.1018601Z * [new branch] gh/PaulZhang12/37/base -> origin/gh/PaulZhang12/37/base 2025-12-04T09:16:47.1020746Z * [new branch] gh/PaulZhang12/37/head -> origin/gh/PaulZhang12/37/head 2025-12-04T09:16:47.1023071Z * [new branch] gh/PaulZhang12/37/orig -> origin/gh/PaulZhang12/37/orig 2025-12-04T09:16:47.1026127Z * [new branch] gh/PaulZhang12/40/base -> origin/gh/PaulZhang12/40/base 2025-12-04T09:16:47.1028369Z * [new branch] gh/PaulZhang12/40/head -> origin/gh/PaulZhang12/40/head 2025-12-04T09:16:47.1030629Z * [new branch] gh/PaulZhang12/40/orig -> origin/gh/PaulZhang12/40/orig 2025-12-04T09:16:47.1033628Z * [new branch] gh/PaulZhang12/42/base -> origin/gh/PaulZhang12/42/base 2025-12-04T09:16:47.1036199Z * [new branch] gh/PaulZhang12/42/head -> origin/gh/PaulZhang12/42/head 2025-12-04T09:16:47.1039343Z * [new branch] gh/PaulZhang12/43/base -> origin/gh/PaulZhang12/43/base 2025-12-04T09:16:47.1041551Z * [new branch] gh/PaulZhang12/43/head -> origin/gh/PaulZhang12/43/head 2025-12-04T09:16:47.1043846Z * [new branch] gh/PaulZhang12/43/orig -> origin/gh/PaulZhang12/43/orig 2025-12-04T09:16:47.1046726Z * [new branch] gh/PaulZhang12/44/base -> origin/gh/PaulZhang12/44/base 2025-12-04T09:16:47.1049167Z * [new branch] gh/PaulZhang12/44/head -> origin/gh/PaulZhang12/44/head 2025-12-04T09:16:47.1052268Z * [new branch] gh/PaulZhang12/45/base -> origin/gh/PaulZhang12/45/base 2025-12-04T09:16:47.1054408Z * [new branch] gh/PaulZhang12/45/head -> origin/gh/PaulZhang12/45/head 2025-12-04T09:16:47.1056611Z * [new branch] gh/PaulZhang12/45/orig -> origin/gh/PaulZhang12/45/orig 2025-12-04T09:16:47.1059730Z * [new branch] gh/PaulZhang12/46/base -> origin/gh/PaulZhang12/46/base 2025-12-04T09:16:47.1062003Z * [new branch] gh/PaulZhang12/46/head -> origin/gh/PaulZhang12/46/head 2025-12-04T09:16:47.1064370Z * [new branch] gh/PaulZhang12/46/orig -> origin/gh/PaulZhang12/46/orig 2025-12-04T09:16:47.1067392Z * [new branch] gh/PaulZhang12/47/base -> origin/gh/PaulZhang12/47/base 2025-12-04T09:16:47.1069679Z * [new branch] gh/PaulZhang12/47/head -> origin/gh/PaulZhang12/47/head 2025-12-04T09:16:47.1071967Z * [new branch] gh/PaulZhang12/47/orig -> origin/gh/PaulZhang12/47/orig 2025-12-04T09:16:47.1074826Z * [new branch] gh/PaulZhang12/48/base -> origin/gh/PaulZhang12/48/base 2025-12-04T09:16:47.1077158Z * [new branch] gh/PaulZhang12/48/head -> origin/gh/PaulZhang12/48/head 2025-12-04T09:16:47.1079414Z * [new branch] gh/PaulZhang12/48/orig -> origin/gh/PaulZhang12/48/orig 2025-12-04T09:16:47.1082924Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-12-04T09:16:47.1085139Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-12-04T09:16:47.1088963Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-12-04T09:16:47.1091314Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-12-04T09:16:47.1094308Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-12-04T09:16:47.1096636Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-12-04T09:16:47.1099041Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-12-04T09:16:47.1102019Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-12-04T09:16:47.1104216Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-12-04T09:16:47.1106763Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-12-04T09:16:47.1109442Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-12-04T09:16:47.1111717Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-12-04T09:16:47.1114023Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-12-04T09:16:47.1117039Z * [new branch] gh/SherlockNoMad/15/base -> origin/gh/SherlockNoMad/15/base 2025-12-04T09:16:47.1119456Z * [new branch] gh/SherlockNoMad/15/head -> origin/gh/SherlockNoMad/15/head 2025-12-04T09:16:47.1121800Z * [new branch] gh/SherlockNoMad/15/orig -> origin/gh/SherlockNoMad/15/orig 2025-12-04T09:16:47.1124768Z * [new branch] gh/SherlockNoMad/17/base -> origin/gh/SherlockNoMad/17/base 2025-12-04T09:16:47.1127078Z * [new branch] gh/SherlockNoMad/17/head -> origin/gh/SherlockNoMad/17/head 2025-12-04T09:16:47.1129453Z * [new branch] gh/SherlockNoMad/17/orig -> origin/gh/SherlockNoMad/17/orig 2025-12-04T09:16:47.1132581Z * [new branch] gh/SherlockNoMad/18/base -> origin/gh/SherlockNoMad/18/base 2025-12-04T09:16:47.1134841Z * [new branch] gh/SherlockNoMad/18/head -> origin/gh/SherlockNoMad/18/head 2025-12-04T09:16:47.1137397Z * [new branch] gh/SherlockNoMad/18/orig -> origin/gh/SherlockNoMad/18/orig 2025-12-04T09:16:47.1140150Z * [new branch] gh/SherlockNoMad/19/base -> origin/gh/SherlockNoMad/19/base 2025-12-04T09:16:47.1142501Z * [new branch] gh/SherlockNoMad/19/head -> origin/gh/SherlockNoMad/19/head 2025-12-04T09:16:47.1144807Z * [new branch] gh/SherlockNoMad/19/orig -> origin/gh/SherlockNoMad/19/orig 2025-12-04T09:16:47.1147668Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-12-04T09:16:47.1149933Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-12-04T09:16:47.1152707Z * [new branch] gh/SherlockNoMad/20/base -> origin/gh/SherlockNoMad/20/base 2025-12-04T09:16:47.1155069Z * [new branch] gh/SherlockNoMad/20/head -> origin/gh/SherlockNoMad/20/head 2025-12-04T09:16:47.1157263Z * [new branch] gh/SherlockNoMad/20/orig -> origin/gh/SherlockNoMad/20/orig 2025-12-04T09:16:47.1160489Z * [new branch] gh/SherlockNoMad/21/base -> origin/gh/SherlockNoMad/21/base 2025-12-04T09:16:47.1162835Z * [new branch] gh/SherlockNoMad/21/head -> origin/gh/SherlockNoMad/21/head 2025-12-04T09:16:47.1165027Z * [new branch] gh/SherlockNoMad/21/orig -> origin/gh/SherlockNoMad/21/orig 2025-12-04T09:16:47.1167978Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-12-04T09:16:47.1170247Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-12-04T09:16:47.1173090Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-12-04T09:16:47.1175229Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-12-04T09:16:47.1178111Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-12-04T09:16:47.1180415Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-12-04T09:16:47.1184562Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-12-04T09:16:47.1187394Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-12-04T09:16:47.1190150Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-12-04T09:16:47.1193181Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-12-04T09:16:47.1196944Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-12-04T09:16:47.1199033Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-12-04T09:16:47.1202147Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-12-04T09:16:47.1204374Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-12-04T09:16:47.1207257Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-12-04T09:16:47.1209661Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-12-04T09:16:47.1212648Z * [new branch] gh/StrongerXi/73/base -> origin/gh/StrongerXi/73/base 2025-12-04T09:16:47.1214948Z * [new branch] gh/StrongerXi/73/head -> origin/gh/StrongerXi/73/head 2025-12-04T09:16:47.1217193Z * [new branch] gh/StrongerXi/73/orig -> origin/gh/StrongerXi/73/orig 2025-12-04T09:16:47.1220845Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-12-04T09:16:47.1223026Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-12-04T09:16:47.1225277Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-12-04T09:16:47.1228291Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-12-04T09:16:47.1230620Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-12-04T09:16:47.1232890Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-12-04T09:16:47.1236312Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-12-04T09:16:47.1238562Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-12-04T09:16:47.1240800Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-12-04T09:16:47.1243831Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-12-04T09:16:47.1246028Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-12-04T09:16:47.1248510Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-12-04T09:16:47.1251408Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-12-04T09:16:47.1253614Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-12-04T09:16:47.1255852Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-12-04T09:16:47.1259024Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-12-04T09:16:47.1261267Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-12-04T09:16:47.1263635Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-12-04T09:16:47.1266543Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-12-04T09:16:47.1268917Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-12-04T09:16:47.1271167Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-12-04T09:16:47.1274147Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-12-04T09:16:47.1276444Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-12-04T09:16:47.1278720Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-12-04T09:16:47.1281919Z * [new branch] gh/XilunWu/176/base -> origin/gh/XilunWu/176/base 2025-12-04T09:16:47.1284149Z * [new branch] gh/XilunWu/176/head -> origin/gh/XilunWu/176/head 2025-12-04T09:16:47.1286616Z * [new branch] gh/XilunWu/176/orig -> origin/gh/XilunWu/176/orig 2025-12-04T09:16:47.1290200Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-12-04T09:16:47.1292443Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-12-04T09:16:47.1294713Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-12-04T09:16:47.1297731Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-12-04T09:16:47.1300012Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-12-04T09:16:47.1302405Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-12-04T09:16:47.1305319Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-12-04T09:16:47.1307558Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-12-04T09:16:47.1309870Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-12-04T09:16:47.1312936Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-12-04T09:16:47.1315234Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-12-04T09:16:47.1317509Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-12-04T09:16:47.1320463Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-12-04T09:16:47.1322815Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-12-04T09:16:47.1325117Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-12-04T09:16:47.1328107Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-12-04T09:16:47.1330462Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-12-04T09:16:47.1332775Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-12-04T09:16:47.1335860Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-12-04T09:16:47.1340664Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-12-04T09:16:47.1342890Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-12-04T09:16:47.1345959Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-12-04T09:16:47.1348267Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-12-04T09:16:47.1350475Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-12-04T09:16:47.1353568Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-12-04T09:16:47.1355860Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-12-04T09:16:47.1358109Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-12-04T09:16:47.1361072Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-12-04T09:16:47.1363326Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-12-04T09:16:47.1365706Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-12-04T09:16:47.1368823Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-12-04T09:16:47.1371074Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-12-04T09:16:47.1373325Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-12-04T09:16:47.1376588Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-12-04T09:16:47.1378629Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-12-04T09:16:47.1380880Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-12-04T09:16:47.1383927Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-12-04T09:16:47.1386522Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-12-04T09:16:47.1389513Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-12-04T09:16:47.1391721Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-12-04T09:16:47.1394083Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-12-04T09:16:47.1397206Z * [new branch] gh/XuehaiPan/390/base -> origin/gh/XuehaiPan/390/base 2025-12-04T09:16:47.1399589Z * [new branch] gh/XuehaiPan/390/head -> origin/gh/XuehaiPan/390/head 2025-12-04T09:16:47.1401808Z * [new branch] gh/XuehaiPan/390/orig -> origin/gh/XuehaiPan/390/orig 2025-12-04T09:16:47.1404770Z * [new branch] gh/XuehaiPan/391/base -> origin/gh/XuehaiPan/391/base 2025-12-04T09:16:47.1406996Z * [new branch] gh/XuehaiPan/391/head -> origin/gh/XuehaiPan/391/head 2025-12-04T09:16:47.1409484Z * [new branch] gh/XuehaiPan/391/orig -> origin/gh/XuehaiPan/391/orig 2025-12-04T09:16:47.1412395Z * [new branch] gh/XuehaiPan/392/base -> origin/gh/XuehaiPan/392/base 2025-12-04T09:16:47.1414637Z * [new branch] gh/XuehaiPan/392/head -> origin/gh/XuehaiPan/392/head 2025-12-04T09:16:47.1416927Z * [new branch] gh/XuehaiPan/392/orig -> origin/gh/XuehaiPan/392/orig 2025-12-04T09:16:47.1420477Z * [new branch] gh/XuehaiPan/394/base -> origin/gh/XuehaiPan/394/base 2025-12-04T09:16:47.1422731Z * [new branch] gh/XuehaiPan/394/head -> origin/gh/XuehaiPan/394/head 2025-12-04T09:16:47.1425097Z * [new branch] gh/XuehaiPan/394/orig -> origin/gh/XuehaiPan/394/orig 2025-12-04T09:16:47.1428248Z * [new branch] gh/XuehaiPan/397/base -> origin/gh/XuehaiPan/397/base 2025-12-04T09:16:47.1430543Z * [new branch] gh/XuehaiPan/397/head -> origin/gh/XuehaiPan/397/head 2025-12-04T09:16:47.1432811Z * [new branch] gh/XuehaiPan/397/orig -> origin/gh/XuehaiPan/397/orig 2025-12-04T09:16:47.1436095Z * [new branch] gh/XuehaiPan/398/base -> origin/gh/XuehaiPan/398/base 2025-12-04T09:16:47.1438404Z * [new branch] gh/XuehaiPan/398/head -> origin/gh/XuehaiPan/398/head 2025-12-04T09:16:47.1440482Z * [new branch] gh/XuehaiPan/398/orig -> origin/gh/XuehaiPan/398/orig 2025-12-04T09:16:47.1443054Z * [new branch] gh/XuehaiPan/399/base -> origin/gh/XuehaiPan/399/base 2025-12-04T09:16:47.1444811Z * [new branch] gh/XuehaiPan/399/head -> origin/gh/XuehaiPan/399/head 2025-12-04T09:16:47.1446680Z * [new branch] gh/XuehaiPan/399/orig -> origin/gh/XuehaiPan/399/orig 2025-12-04T09:16:47.1449346Z * [new branch] gh/XuehaiPan/400/base -> origin/gh/XuehaiPan/400/base 2025-12-04T09:16:47.1451055Z * [new branch] gh/XuehaiPan/400/head -> origin/gh/XuehaiPan/400/head 2025-12-04T09:16:47.1452883Z * [new branch] gh/XuehaiPan/400/orig -> origin/gh/XuehaiPan/400/orig 2025-12-04T09:16:47.1456185Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-12-04T09:16:47.1458041Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-12-04T09:16:47.1460064Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-12-04T09:16:47.1463215Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-12-04T09:16:47.1464958Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-12-04T09:16:47.1467457Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-12-04T09:16:47.1469359Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-12-04T09:16:47.1472165Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-12-04T09:16:47.1474056Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-12-04T09:16:47.1476596Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-12-04T09:16:47.1478577Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-12-04T09:16:47.1481198Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-12-04T09:16:47.1483070Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-12-04T09:16:47.1485802Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-12-04T09:16:47.1487693Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-12-04T09:16:47.1490340Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-12-04T09:16:47.1492116Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-12-04T09:16:47.1494083Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-12-04T09:16:47.1497441Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-12-04T09:16:47.1499415Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-12-04T09:16:47.1502177Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-12-04T09:16:47.1504170Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-12-04T09:16:47.1506740Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-12-04T09:16:47.1508573Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-12-04T09:16:47.1510475Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-12-04T09:16:47.1513985Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-12-04T09:16:47.1515849Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-12-04T09:16:47.1518282Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-12-04T09:16:47.1521312Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-12-04T09:16:47.1524367Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-12-04T09:16:47.1526229Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-12-04T09:16:47.1528360Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-12-04T09:16:47.1531218Z * [new branch] gh/alexsamardzic/14/base -> origin/gh/alexsamardzic/14/base 2025-12-04T09:16:47.1533092Z * [new branch] gh/alexsamardzic/14/head -> origin/gh/alexsamardzic/14/head 2025-12-04T09:16:47.1534944Z * [new branch] gh/alexsamardzic/14/orig -> origin/gh/alexsamardzic/14/orig 2025-12-04T09:16:47.1537821Z * [new branch] gh/alexsamardzic/15/base -> origin/gh/alexsamardzic/15/base 2025-12-04T09:16:47.1539994Z * [new branch] gh/alexsamardzic/15/head -> origin/gh/alexsamardzic/15/head 2025-12-04T09:16:47.1542021Z * [new branch] gh/alexsamardzic/15/orig -> origin/gh/alexsamardzic/15/orig 2025-12-04T09:16:47.1545033Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-12-04T09:16:47.1546877Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-12-04T09:16:47.1549297Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-12-04T09:16:47.1552634Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-12-04T09:16:47.1554591Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-12-04T09:16:47.1556830Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-12-04T09:16:47.1559516Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-12-04T09:16:47.1561500Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-12-04T09:16:47.1563484Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-12-04T09:16:47.1566910Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-12-04T09:16:47.1569310Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-12-04T09:16:47.1571903Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-12-04T09:16:47.1574337Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-12-04T09:16:47.1577060Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-12-04T09:16:47.1579029Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-12-04T09:16:47.1581227Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-12-04T09:16:47.1583990Z * [new branch] gh/andyanwang/42/base -> origin/gh/andyanwang/42/base 2025-12-04T09:16:47.1585790Z * [new branch] gh/andyanwang/42/head -> origin/gh/andyanwang/42/head 2025-12-04T09:16:47.1587663Z * [new branch] gh/andyanwang/42/orig -> origin/gh/andyanwang/42/orig 2025-12-04T09:16:47.1590654Z * [new branch] gh/andyanwang/45/base -> origin/gh/andyanwang/45/base 2025-12-04T09:16:47.1592619Z * [new branch] gh/andyanwang/45/head -> origin/gh/andyanwang/45/head 2025-12-04T09:16:47.1594505Z * [new branch] gh/andyanwang/45/orig -> origin/gh/andyanwang/45/orig 2025-12-04T09:16:47.1597584Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-12-04T09:16:47.1599607Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-12-04T09:16:47.1602305Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-12-04T09:16:47.1604246Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-12-04T09:16:47.1606262Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-12-04T09:16:47.1609341Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-12-04T09:16:47.1611205Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-12-04T09:16:47.1613007Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-12-04T09:16:47.1615902Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-12-04T09:16:47.1617686Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-12-04T09:16:47.1619578Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-12-04T09:16:47.1622511Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-12-04T09:16:47.1625010Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-12-04T09:16:47.1627284Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-12-04T09:16:47.1629847Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-12-04T09:16:47.1631702Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-12-04T09:16:47.1633499Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-12-04T09:16:47.1636129Z * [new branch] gh/angelayi/131/base -> origin/gh/angelayi/131/base 2025-12-04T09:16:47.1638462Z * [new branch] gh/angelayi/131/head -> origin/gh/angelayi/131/head 2025-12-04T09:16:47.1640291Z * [new branch] gh/angelayi/131/orig -> origin/gh/angelayi/131/orig 2025-12-04T09:16:47.1643155Z * [new branch] gh/angelayi/132/base -> origin/gh/angelayi/132/base 2025-12-04T09:16:47.1645161Z * [new branch] gh/angelayi/132/head -> origin/gh/angelayi/132/head 2025-12-04T09:16:47.1647101Z * [new branch] gh/angelayi/132/orig -> origin/gh/angelayi/132/orig 2025-12-04T09:16:47.1649912Z * [new branch] gh/angelayi/133/base -> origin/gh/angelayi/133/base 2025-12-04T09:16:47.1651760Z * [new branch] gh/angelayi/133/head -> origin/gh/angelayi/133/head 2025-12-04T09:16:47.1653635Z * [new branch] gh/angelayi/133/orig -> origin/gh/angelayi/133/orig 2025-12-04T09:16:47.1656534Z * [new branch] gh/angelayi/134/base -> origin/gh/angelayi/134/base 2025-12-04T09:16:47.1658466Z * [new branch] gh/angelayi/134/head -> origin/gh/angelayi/134/head 2025-12-04T09:16:47.1660309Z * [new branch] gh/angelayi/134/orig -> origin/gh/angelayi/134/orig 2025-12-04T09:16:47.1663016Z * [new branch] gh/angelayi/135/base -> origin/gh/angelayi/135/base 2025-12-04T09:16:47.1664911Z * [new branch] gh/angelayi/135/head -> origin/gh/angelayi/135/head 2025-12-04T09:16:47.1666760Z * [new branch] gh/angelayi/135/orig -> origin/gh/angelayi/135/orig 2025-12-04T09:16:47.1669328Z * [new branch] gh/angelayi/136/base -> origin/gh/angelayi/136/base 2025-12-04T09:16:47.1671130Z * [new branch] gh/angelayi/136/head -> origin/gh/angelayi/136/head 2025-12-04T09:16:47.1672948Z * [new branch] gh/angelayi/136/orig -> origin/gh/angelayi/136/orig 2025-12-04T09:16:47.1675504Z * [new branch] gh/angelayi/137/base -> origin/gh/angelayi/137/base 2025-12-04T09:16:47.1677252Z * [new branch] gh/angelayi/137/head -> origin/gh/angelayi/137/head 2025-12-04T09:16:47.1679277Z * [new branch] gh/angelayi/137/orig -> origin/gh/angelayi/137/orig 2025-12-04T09:16:47.1681773Z * [new branch] gh/angelayi/138/base -> origin/gh/angelayi/138/base 2025-12-04T09:16:47.1683525Z * [new branch] gh/angelayi/138/head -> origin/gh/angelayi/138/head 2025-12-04T09:16:47.1685317Z * [new branch] gh/angelayi/138/orig -> origin/gh/angelayi/138/orig 2025-12-04T09:16:47.1687908Z * [new branch] gh/angelayi/139/base -> origin/gh/angelayi/139/base 2025-12-04T09:16:47.1689788Z * [new branch] gh/angelayi/139/head -> origin/gh/angelayi/139/head 2025-12-04T09:16:47.1691582Z * [new branch] gh/angelayi/139/orig -> origin/gh/angelayi/139/orig 2025-12-04T09:16:47.1694230Z * [new branch] gh/angelayi/140/base -> origin/gh/angelayi/140/base 2025-12-04T09:16:47.1696114Z * [new branch] gh/angelayi/140/head -> origin/gh/angelayi/140/head 2025-12-04T09:16:47.1697997Z * [new branch] gh/angelayi/140/orig -> origin/gh/angelayi/140/orig 2025-12-04T09:16:47.1701336Z * [new branch] gh/angelayi/141/base -> origin/gh/angelayi/141/base 2025-12-04T09:16:47.1703144Z * [new branch] gh/angelayi/141/head -> origin/gh/angelayi/141/head 2025-12-04T09:16:47.1704971Z * [new branch] gh/angelayi/141/orig -> origin/gh/angelayi/141/orig 2025-12-04T09:16:47.1707691Z * [new branch] gh/angelayi/142/base -> origin/gh/angelayi/142/base 2025-12-04T09:16:47.1709558Z * [new branch] gh/angelayi/142/head -> origin/gh/angelayi/142/head 2025-12-04T09:16:47.1711333Z * [new branch] gh/angelayi/142/orig -> origin/gh/angelayi/142/orig 2025-12-04T09:16:47.1713899Z * [new branch] gh/angelayi/143/base -> origin/gh/angelayi/143/base 2025-12-04T09:16:47.1715724Z * [new branch] gh/angelayi/143/head -> origin/gh/angelayi/143/head 2025-12-04T09:16:47.1717553Z * [new branch] gh/angelayi/143/orig -> origin/gh/angelayi/143/orig 2025-12-04T09:16:47.1720179Z * [new branch] gh/angelayi/144/base -> origin/gh/angelayi/144/base 2025-12-04T09:16:47.1722113Z * [new branch] gh/angelayi/144/head -> origin/gh/angelayi/144/head 2025-12-04T09:16:47.1723961Z * [new branch] gh/angelayi/144/orig -> origin/gh/angelayi/144/orig 2025-12-04T09:16:47.1727182Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-12-04T09:16:47.1729189Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-12-04T09:16:47.1730956Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-12-04T09:16:47.1733745Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-12-04T09:16:47.1735739Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-12-04T09:16:47.1739797Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-12-04T09:16:47.1742485Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-12-04T09:16:47.1744488Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-12-04T09:16:47.1746562Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-12-04T09:16:47.1749119Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-12-04T09:16:47.1750446Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-12-04T09:16:47.1752791Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-12-04T09:16:47.1755366Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-12-04T09:16:47.1756638Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-12-04T09:16:47.1758906Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-12-04T09:16:47.1761553Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-12-04T09:16:47.1762848Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-12-04T09:16:47.1765020Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-12-04T09:16:47.1767779Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-12-04T09:16:47.1769922Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-12-04T09:16:47.1771112Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-12-04T09:16:47.1774309Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-12-04T09:16:47.1775563Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-12-04T09:16:47.1778009Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-12-04T09:16:47.1780418Z * [new branch] gh/anijain2305/910/base -> origin/gh/anijain2305/910/base 2025-12-04T09:16:47.1782274Z * [new branch] gh/anijain2305/910/head -> origin/gh/anijain2305/910/head 2025-12-04T09:16:47.1783737Z * [new branch] gh/anijain2305/910/orig -> origin/gh/anijain2305/910/orig 2025-12-04T09:16:47.1786901Z * [new branch] gh/anijain2305/919/base -> origin/gh/anijain2305/919/base 2025-12-04T09:16:47.1788555Z * [new branch] gh/anijain2305/919/head -> origin/gh/anijain2305/919/head 2025-12-04T09:16:47.1790807Z * [new branch] gh/anijain2305/919/orig -> origin/gh/anijain2305/919/orig 2025-12-04T09:16:47.1793208Z * [new branch] gh/anijain2305/922/base -> origin/gh/anijain2305/922/base 2025-12-04T09:16:47.1794860Z * [new branch] gh/anijain2305/922/head -> origin/gh/anijain2305/922/head 2025-12-04T09:16:47.1796870Z * [new branch] gh/anijain2305/922/orig -> origin/gh/anijain2305/922/orig 2025-12-04T09:16:47.1799528Z * [new branch] gh/anijain2305/932/base -> origin/gh/anijain2305/932/base 2025-12-04T09:16:47.1801510Z * [new branch] gh/anijain2305/932/head -> origin/gh/anijain2305/932/head 2025-12-04T09:16:47.1802876Z * [new branch] gh/anijain2305/932/orig -> origin/gh/anijain2305/932/orig 2025-12-04T09:16:47.1810960Z * [new branch] gh/anijain2305/940/base -> origin/gh/anijain2305/940/base 2025-12-04T09:16:47.1811559Z * [new branch] gh/anijain2305/940/head -> origin/gh/anijain2305/940/head 2025-12-04T09:16:47.1812110Z * [new branch] gh/anijain2305/940/orig -> origin/gh/anijain2305/940/orig 2025-12-04T09:16:47.1812677Z * [new branch] gh/anijain2305/941/base -> origin/gh/anijain2305/941/base 2025-12-04T09:16:47.1813834Z * [new branch] gh/anijain2305/941/head -> origin/gh/anijain2305/941/head 2025-12-04T09:16:47.1815786Z * [new branch] gh/anijain2305/941/orig -> origin/gh/anijain2305/941/orig 2025-12-04T09:16:47.1818278Z * [new branch] gh/anijain2305/942/base -> origin/gh/anijain2305/942/base 2025-12-04T09:16:47.1820189Z * [new branch] gh/anijain2305/942/head -> origin/gh/anijain2305/942/head 2025-12-04T09:16:47.1822063Z * [new branch] gh/anijain2305/942/orig -> origin/gh/anijain2305/942/orig 2025-12-04T09:16:47.1824697Z * [new branch] gh/anijain2305/943/base -> origin/gh/anijain2305/943/base 2025-12-04T09:16:47.1826566Z * [new branch] gh/anijain2305/943/head -> origin/gh/anijain2305/943/head 2025-12-04T09:16:47.1828291Z * [new branch] gh/anijain2305/943/orig -> origin/gh/anijain2305/943/orig 2025-12-04T09:16:47.1831498Z * [new branch] gh/anijain2305/944/base -> origin/gh/anijain2305/944/base 2025-12-04T09:16:47.1833327Z * [new branch] gh/anijain2305/944/head -> origin/gh/anijain2305/944/head 2025-12-04T09:16:47.1835871Z * [new branch] gh/anijain2305/944/orig -> origin/gh/anijain2305/944/orig 2025-12-04T09:16:47.1838624Z * [new branch] gh/anijain2305/945/base -> origin/gh/anijain2305/945/base 2025-12-04T09:16:47.1840417Z * [new branch] gh/anijain2305/945/head -> origin/gh/anijain2305/945/head 2025-12-04T09:16:47.1842239Z * [new branch] gh/anijain2305/945/orig -> origin/gh/anijain2305/945/orig 2025-12-04T09:16:47.1846297Z * [new branch] gh/anijain2305/946/base -> origin/gh/anijain2305/946/base 2025-12-04T09:16:47.1848829Z * [new branch] gh/anijain2305/946/head -> origin/gh/anijain2305/946/head 2025-12-04T09:16:47.1851523Z * [new branch] gh/anijain2305/946/orig -> origin/gh/anijain2305/946/orig 2025-12-04T09:16:47.1854369Z * [new branch] gh/anijain2305/947/base -> origin/gh/anijain2305/947/base 2025-12-04T09:16:47.1855727Z * [new branch] gh/anijain2305/947/head -> origin/gh/anijain2305/947/head 2025-12-04T09:16:47.1857779Z * [new branch] gh/anijain2305/947/orig -> origin/gh/anijain2305/947/orig 2025-12-04T09:16:47.1860496Z * [new branch] gh/anijain2305/948/base -> origin/gh/anijain2305/948/base 2025-12-04T09:16:47.1861961Z * [new branch] gh/anijain2305/948/head -> origin/gh/anijain2305/948/head 2025-12-04T09:16:47.1864036Z * [new branch] gh/anijain2305/948/orig -> origin/gh/anijain2305/948/orig 2025-12-04T09:16:47.1866768Z * [new branch] gh/anijain2305/949/base -> origin/gh/anijain2305/949/base 2025-12-04T09:16:47.1868296Z * [new branch] gh/anijain2305/949/head -> origin/gh/anijain2305/949/head 2025-12-04T09:16:47.1870400Z * [new branch] gh/anijain2305/949/orig -> origin/gh/anijain2305/949/orig 2025-12-04T09:16:47.1873028Z * [new branch] gh/anijain2305/950/base -> origin/gh/anijain2305/950/base 2025-12-04T09:16:47.1874651Z * [new branch] gh/anijain2305/950/head -> origin/gh/anijain2305/950/head 2025-12-04T09:16:47.1876640Z * [new branch] gh/anijain2305/950/orig -> origin/gh/anijain2305/950/orig 2025-12-04T09:16:47.1879285Z * [new branch] gh/anijain2305/951/base -> origin/gh/anijain2305/951/base 2025-12-04T09:16:47.1880840Z * [new branch] gh/anijain2305/951/head -> origin/gh/anijain2305/951/head 2025-12-04T09:16:47.1883031Z * [new branch] gh/anijain2305/951/orig -> origin/gh/anijain2305/951/orig 2025-12-04T09:16:47.1885672Z * [new branch] gh/anijain2305/952/base -> origin/gh/anijain2305/952/base 2025-12-04T09:16:47.1887151Z * [new branch] gh/anijain2305/952/head -> origin/gh/anijain2305/952/head 2025-12-04T09:16:47.1889377Z * [new branch] gh/anijain2305/952/orig -> origin/gh/anijain2305/952/orig 2025-12-04T09:16:47.1892133Z * [new branch] gh/anijain2305/953/base -> origin/gh/anijain2305/953/base 2025-12-04T09:16:47.1893684Z * [new branch] gh/anijain2305/953/head -> origin/gh/anijain2305/953/head 2025-12-04T09:16:47.1895773Z * [new branch] gh/anijain2305/953/orig -> origin/gh/anijain2305/953/orig 2025-12-04T09:16:47.1898525Z * [new branch] gh/anijain2305/954/base -> origin/gh/anijain2305/954/base 2025-12-04T09:16:47.1900409Z * [new branch] gh/anijain2305/954/head -> origin/gh/anijain2305/954/head 2025-12-04T09:16:47.1902569Z * [new branch] gh/anijain2305/954/orig -> origin/gh/anijain2305/954/orig 2025-12-04T09:16:47.1904989Z * [new branch] gh/anijain2305/955/base -> origin/gh/anijain2305/955/base 2025-12-04T09:16:47.1906558Z * [new branch] gh/anijain2305/955/head -> origin/gh/anijain2305/955/head 2025-12-04T09:16:47.1908658Z * [new branch] gh/anijain2305/955/orig -> origin/gh/anijain2305/955/orig 2025-12-04T09:16:47.1911336Z * [new branch] gh/anijain2305/956/base -> origin/gh/anijain2305/956/base 2025-12-04T09:16:47.1913220Z * [new branch] gh/anijain2305/956/head -> origin/gh/anijain2305/956/head 2025-12-04T09:16:47.1914786Z * [new branch] gh/anijain2305/956/orig -> origin/gh/anijain2305/956/orig 2025-12-04T09:16:47.1917870Z * [new branch] gh/anijain2305/957/base -> origin/gh/anijain2305/957/base 2025-12-04T09:16:47.1919740Z * [new branch] gh/anijain2305/957/head -> origin/gh/anijain2305/957/head 2025-12-04T09:16:47.1921658Z * [new branch] gh/anijain2305/957/orig -> origin/gh/anijain2305/957/orig 2025-12-04T09:16:47.1924262Z * [new branch] gh/anijain2305/958/base -> origin/gh/anijain2305/958/base 2025-12-04T09:16:47.1926920Z * [new branch] gh/anijain2305/958/head -> origin/gh/anijain2305/958/head 2025-12-04T09:16:47.1928441Z * [new branch] gh/anijain2305/958/orig -> origin/gh/anijain2305/958/orig 2025-12-04T09:16:47.1931374Z * [new branch] gh/anijain2305/959/base -> origin/gh/anijain2305/959/base 2025-12-04T09:16:47.1932950Z * [new branch] gh/anijain2305/959/head -> origin/gh/anijain2305/959/head 2025-12-04T09:16:47.1934990Z * [new branch] gh/anijain2305/959/orig -> origin/gh/anijain2305/959/orig 2025-12-04T09:16:47.1937884Z * [new branch] gh/anijain2305/960/base -> origin/gh/anijain2305/960/base 2025-12-04T09:16:47.1939761Z * [new branch] gh/anijain2305/960/head -> origin/gh/anijain2305/960/head 2025-12-04T09:16:47.1941610Z * [new branch] gh/anijain2305/960/orig -> origin/gh/anijain2305/960/orig 2025-12-04T09:16:47.1944362Z * [new branch] gh/anijain2305/961/base -> origin/gh/anijain2305/961/base 2025-12-04T09:16:47.1946311Z * [new branch] gh/anijain2305/961/head -> origin/gh/anijain2305/961/head 2025-12-04T09:16:47.1947937Z * [new branch] gh/anijain2305/961/orig -> origin/gh/anijain2305/961/orig 2025-12-04T09:16:47.1950695Z * [new branch] gh/anijain2305/962/base -> origin/gh/anijain2305/962/base 2025-12-04T09:16:47.1952563Z * [new branch] gh/anijain2305/962/head -> origin/gh/anijain2305/962/head 2025-12-04T09:16:47.1954199Z * [new branch] gh/anijain2305/962/orig -> origin/gh/anijain2305/962/orig 2025-12-04T09:16:47.1957295Z * [new branch] gh/anijain2305/963/base -> origin/gh/anijain2305/963/base 2025-12-04T09:16:47.1959459Z * [new branch] gh/anijain2305/963/head -> origin/gh/anijain2305/963/head 2025-12-04T09:16:47.1961369Z * [new branch] gh/anijain2305/963/orig -> origin/gh/anijain2305/963/orig 2025-12-04T09:16:47.1963915Z * [new branch] gh/anijain2305/964/base -> origin/gh/anijain2305/964/base 2025-12-04T09:16:47.1965793Z * [new branch] gh/anijain2305/964/head -> origin/gh/anijain2305/964/head 2025-12-04T09:16:47.1967597Z * [new branch] gh/anijain2305/964/orig -> origin/gh/anijain2305/964/orig 2025-12-04T09:16:47.1970585Z * [new branch] gh/anijain2305/965/base -> origin/gh/anijain2305/965/base 2025-12-04T09:16:47.1972492Z * [new branch] gh/anijain2305/965/head -> origin/gh/anijain2305/965/head 2025-12-04T09:16:47.1974471Z * [new branch] gh/anijain2305/965/orig -> origin/gh/anijain2305/965/orig 2025-12-04T09:16:47.1976924Z * [new branch] gh/anijain2305/966/base -> origin/gh/anijain2305/966/base 2025-12-04T09:16:47.1978852Z * [new branch] gh/anijain2305/966/head -> origin/gh/anijain2305/966/head 2025-12-04T09:16:47.1980812Z * [new branch] gh/anijain2305/966/orig -> origin/gh/anijain2305/966/orig 2025-12-04T09:16:47.1983354Z * [new branch] gh/anijain2305/967/base -> origin/gh/anijain2305/967/base 2025-12-04T09:16:47.1985184Z * [new branch] gh/anijain2305/967/head -> origin/gh/anijain2305/967/head 2025-12-04T09:16:47.1987196Z * [new branch] gh/anijain2305/967/orig -> origin/gh/anijain2305/967/orig 2025-12-04T09:16:47.1989790Z * [new branch] gh/anijain2305/968/base -> origin/gh/anijain2305/968/base 2025-12-04T09:16:47.1991632Z * [new branch] gh/anijain2305/968/head -> origin/gh/anijain2305/968/head 2025-12-04T09:16:47.1993521Z * [new branch] gh/anijain2305/968/orig -> origin/gh/anijain2305/968/orig 2025-12-04T09:16:47.1996177Z * [new branch] gh/anijain2305/969/base -> origin/gh/anijain2305/969/base 2025-12-04T09:16:47.1998067Z * [new branch] gh/anijain2305/969/head -> origin/gh/anijain2305/969/head 2025-12-04T09:16:47.1999965Z * [new branch] gh/anijain2305/969/orig -> origin/gh/anijain2305/969/orig 2025-12-04T09:16:47.2002588Z * [new branch] gh/anijain2305/970/base -> origin/gh/anijain2305/970/base 2025-12-04T09:16:47.2004428Z * [new branch] gh/anijain2305/970/head -> origin/gh/anijain2305/970/head 2025-12-04T09:16:47.2006720Z * [new branch] gh/anijain2305/970/orig -> origin/gh/anijain2305/970/orig 2025-12-04T09:16:47.2009705Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-12-04T09:16:47.2011424Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-12-04T09:16:47.2013290Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-12-04T09:16:47.2016558Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-12-04T09:16:47.2018617Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-12-04T09:16:47.2020964Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-12-04T09:16:47.2022836Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-12-04T09:16:47.2024977Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-12-04T09:16:47.2026825Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-12-04T09:16:47.2029160Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-12-04T09:16:47.2030915Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-12-04T09:16:47.2033285Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-12-04T09:16:47.2035153Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-12-04T09:16:47.2038104Z * [new branch] gh/anshul-si/53/base -> origin/gh/anshul-si/53/base 2025-12-04T09:16:47.2039912Z * [new branch] gh/anshul-si/53/head -> origin/gh/anshul-si/53/head 2025-12-04T09:16:47.2042647Z * [new branch] gh/anshul-si/58/base -> origin/gh/anshul-si/58/base 2025-12-04T09:16:47.2044442Z * [new branch] gh/anshul-si/58/head -> origin/gh/anshul-si/58/head 2025-12-04T09:16:47.2046829Z * [new branch] gh/anshul-si/66/base -> origin/gh/anshul-si/66/base 2025-12-04T09:16:47.2048991Z * [new branch] gh/anshul-si/66/head -> origin/gh/anshul-si/66/head 2025-12-04T09:16:47.2050815Z * [new branch] gh/anshul-si/66/orig -> origin/gh/anshul-si/66/orig 2025-12-04T09:16:47.2053173Z * [new branch] gh/anshul-si/67/base -> origin/gh/anshul-si/67/base 2025-12-04T09:16:47.2055015Z * [new branch] gh/anshul-si/67/head -> origin/gh/anshul-si/67/head 2025-12-04T09:16:47.2056844Z * [new branch] gh/anshul-si/67/orig -> origin/gh/anshul-si/67/orig 2025-12-04T09:16:47.2059600Z * [new branch] gh/anshul-si/68/base -> origin/gh/anshul-si/68/base 2025-12-04T09:16:47.2061343Z * [new branch] gh/anshul-si/68/head -> origin/gh/anshul-si/68/head 2025-12-04T09:16:47.2063122Z * [new branch] gh/anshul-si/68/orig -> origin/gh/anshul-si/68/orig 2025-12-04T09:16:47.2065984Z * [new branch] gh/anshul-si/69/base -> origin/gh/anshul-si/69/base 2025-12-04T09:16:47.2067779Z * [new branch] gh/anshul-si/69/head -> origin/gh/anshul-si/69/head 2025-12-04T09:16:47.2069711Z * [new branch] gh/anshul-si/69/orig -> origin/gh/anshul-si/69/orig 2025-12-04T09:16:47.2072196Z * [new branch] gh/anshul-si/70/base -> origin/gh/anshul-si/70/base 2025-12-04T09:16:47.2074460Z * [new branch] gh/anshul-si/70/head -> origin/gh/anshul-si/70/head 2025-12-04T09:16:47.2076564Z * [new branch] gh/anshul-si/70/orig -> origin/gh/anshul-si/70/orig 2025-12-04T09:16:47.2078887Z * [new branch] gh/anshul-si/71/base -> origin/gh/anshul-si/71/base 2025-12-04T09:16:47.2080760Z * [new branch] gh/anshul-si/71/head -> origin/gh/anshul-si/71/head 2025-12-04T09:16:47.2082472Z * [new branch] gh/anshul-si/71/orig -> origin/gh/anshul-si/71/orig 2025-12-04T09:16:47.2085172Z * [new branch] gh/anshul-si/72/base -> origin/gh/anshul-si/72/base 2025-12-04T09:16:47.2087056Z * [new branch] gh/anshul-si/72/head -> origin/gh/anshul-si/72/head 2025-12-04T09:16:47.2089068Z * [new branch] gh/anshul-si/72/orig -> origin/gh/anshul-si/72/orig 2025-12-04T09:16:47.2091731Z * [new branch] gh/anshul-si/73/base -> origin/gh/anshul-si/73/base 2025-12-04T09:16:47.2093566Z * [new branch] gh/anshul-si/73/head -> origin/gh/anshul-si/73/head 2025-12-04T09:16:47.2095410Z * [new branch] gh/anshul-si/73/orig -> origin/gh/anshul-si/73/orig 2025-12-04T09:16:47.2098624Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-12-04T09:16:47.2100462Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-12-04T09:16:47.2103210Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-12-04T09:16:47.2105209Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-12-04T09:16:47.2107075Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-12-04T09:16:47.2109648Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-12-04T09:16:47.2111500Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-12-04T09:16:47.2113375Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-12-04T09:16:47.2116012Z * [new branch] gh/aorenste/141/base -> origin/gh/aorenste/141/base 2025-12-04T09:16:47.2117894Z * [new branch] gh/aorenste/141/head -> origin/gh/aorenste/141/head 2025-12-04T09:16:47.2120685Z * [new branch] gh/aorenste/145/base -> origin/gh/aorenste/145/base 2025-12-04T09:16:47.2122563Z * [new branch] gh/aorenste/145/head -> origin/gh/aorenste/145/head 2025-12-04T09:16:47.2124557Z * [new branch] gh/aorenste/145/orig -> origin/gh/aorenste/145/orig 2025-12-04T09:16:47.2127166Z * [new branch] gh/aorenste/146/base -> origin/gh/aorenste/146/base 2025-12-04T09:16:47.2129261Z * [new branch] gh/aorenste/146/head -> origin/gh/aorenste/146/head 2025-12-04T09:16:47.2131116Z * [new branch] gh/aorenste/146/orig -> origin/gh/aorenste/146/orig 2025-12-04T09:16:47.2133756Z * [new branch] gh/aorenste/147/base -> origin/gh/aorenste/147/base 2025-12-04T09:16:47.2135883Z * [new branch] gh/aorenste/147/head -> origin/gh/aorenste/147/head 2025-12-04T09:16:47.2137765Z * [new branch] gh/aorenste/147/orig -> origin/gh/aorenste/147/orig 2025-12-04T09:16:47.2140276Z * [new branch] gh/aorenste/148/base -> origin/gh/aorenste/148/base 2025-12-04T09:16:47.2142237Z * [new branch] gh/aorenste/148/head -> origin/gh/aorenste/148/head 2025-12-04T09:16:47.2144129Z * [new branch] gh/aorenste/148/orig -> origin/gh/aorenste/148/orig 2025-12-04T09:16:47.2146675Z * [new branch] gh/aorenste/149/base -> origin/gh/aorenste/149/base 2025-12-04T09:16:47.2148559Z * [new branch] gh/aorenste/149/head -> origin/gh/aorenste/149/head 2025-12-04T09:16:47.2150388Z * [new branch] gh/aorenste/149/orig -> origin/gh/aorenste/149/orig 2025-12-04T09:16:47.2153192Z * [new branch] gh/aorenste/150/base -> origin/gh/aorenste/150/base 2025-12-04T09:16:47.2154860Z * [new branch] gh/aorenste/150/head -> origin/gh/aorenste/150/head 2025-12-04T09:16:47.2156695Z * [new branch] gh/aorenste/150/orig -> origin/gh/aorenste/150/orig 2025-12-04T09:16:47.2159156Z * [new branch] gh/aorenste/151/base -> origin/gh/aorenste/151/base 2025-12-04T09:16:47.2160919Z * [new branch] gh/aorenste/151/head -> origin/gh/aorenste/151/head 2025-12-04T09:16:47.2162872Z * [new branch] gh/aorenste/151/orig -> origin/gh/aorenste/151/orig 2025-12-04T09:16:47.2165482Z * [new branch] gh/aorenste/152/base -> origin/gh/aorenste/152/base 2025-12-04T09:16:47.2167348Z * [new branch] gh/aorenste/152/head -> origin/gh/aorenste/152/head 2025-12-04T09:16:47.2169360Z * [new branch] gh/aorenste/152/orig -> origin/gh/aorenste/152/orig 2025-12-04T09:16:47.2171814Z * [new branch] gh/aorenste/153/base -> origin/gh/aorenste/153/base 2025-12-04T09:16:47.2173539Z * [new branch] gh/aorenste/153/head -> origin/gh/aorenste/153/head 2025-12-04T09:16:47.2175368Z * [new branch] gh/aorenste/153/orig -> origin/gh/aorenste/153/orig 2025-12-04T09:16:47.2178026Z * [new branch] gh/aorenste/154/base -> origin/gh/aorenste/154/base 2025-12-04T09:16:47.2180017Z * [new branch] gh/aorenste/154/head -> origin/gh/aorenste/154/head 2025-12-04T09:16:47.2182423Z * [new branch] gh/aorenste/154/orig -> origin/gh/aorenste/154/orig 2025-12-04T09:16:47.2185831Z * [new branch] gh/aorenste/155/base -> origin/gh/aorenste/155/base 2025-12-04T09:16:47.2187755Z * [new branch] gh/aorenste/155/head -> origin/gh/aorenste/155/head 2025-12-04T09:16:47.2189540Z * [new branch] gh/aorenste/155/orig -> origin/gh/aorenste/155/orig 2025-12-04T09:16:47.2191896Z * [new branch] gh/aorenste/156/base -> origin/gh/aorenste/156/base 2025-12-04T09:16:47.2193851Z * [new branch] gh/aorenste/156/head -> origin/gh/aorenste/156/head 2025-12-04T09:16:47.2195594Z * [new branch] gh/aorenste/156/orig -> origin/gh/aorenste/156/orig 2025-12-04T09:16:47.2198502Z * [new branch] gh/aorenste/157/base -> origin/gh/aorenste/157/base 2025-12-04T09:16:47.2200350Z * [new branch] gh/aorenste/157/head -> origin/gh/aorenste/157/head 2025-12-04T09:16:47.2202181Z * [new branch] gh/aorenste/157/orig -> origin/gh/aorenste/157/orig 2025-12-04T09:16:47.2204627Z * [new branch] gh/aorenste/158/base -> origin/gh/aorenste/158/base 2025-12-04T09:16:47.2206574Z * [new branch] gh/aorenste/158/head -> origin/gh/aorenste/158/head 2025-12-04T09:16:47.2208385Z * [new branch] gh/aorenste/158/orig -> origin/gh/aorenste/158/orig 2025-12-04T09:16:47.2210925Z * [new branch] gh/aorenste/159/base -> origin/gh/aorenste/159/base 2025-12-04T09:16:47.2212773Z * [new branch] gh/aorenste/159/head -> origin/gh/aorenste/159/head 2025-12-04T09:16:47.2214565Z * [new branch] gh/aorenste/159/orig -> origin/gh/aorenste/159/orig 2025-12-04T09:16:47.2217826Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-12-04T09:16:47.2219704Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-12-04T09:16:47.2222060Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-12-04T09:16:47.2223883Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-12-04T09:16:47.2225649Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-12-04T09:16:47.2229107Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-12-04T09:16:47.2230874Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-12-04T09:16:47.2232699Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-12-04T09:16:47.2235471Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-12-04T09:16:47.2239276Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-12-04T09:16:47.2241106Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-12-04T09:16:47.2243823Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-12-04T09:16:47.2245586Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-12-04T09:16:47.2247356Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-12-04T09:16:47.2250318Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-12-04T09:16:47.2252202Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-12-04T09:16:47.2254136Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-12-04T09:16:47.2256731Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-12-04T09:16:47.2258520Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-12-04T09:16:47.2260326Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-12-04T09:16:47.2263122Z * [new branch] gh/bdhirsh/675/base -> origin/gh/bdhirsh/675/base 2025-12-04T09:16:47.2265074Z * [new branch] gh/bdhirsh/675/head -> origin/gh/bdhirsh/675/head 2025-12-04T09:16:47.2266955Z * [new branch] gh/bdhirsh/675/orig -> origin/gh/bdhirsh/675/orig 2025-12-04T09:16:47.2269611Z * [new branch] gh/bdhirsh/676/base -> origin/gh/bdhirsh/676/base 2025-12-04T09:16:47.2271511Z * [new branch] gh/bdhirsh/676/head -> origin/gh/bdhirsh/676/head 2025-12-04T09:16:47.2273343Z * [new branch] gh/bdhirsh/676/orig -> origin/gh/bdhirsh/676/orig 2025-12-04T09:16:47.2275862Z * [new branch] gh/bdhirsh/677/base -> origin/gh/bdhirsh/677/base 2025-12-04T09:16:47.2278069Z * [new branch] gh/bdhirsh/677/head -> origin/gh/bdhirsh/677/head 2025-12-04T09:16:47.2279960Z * [new branch] gh/bdhirsh/677/orig -> origin/gh/bdhirsh/677/orig 2025-12-04T09:16:47.2282551Z * [new branch] gh/bdhirsh/678/base -> origin/gh/bdhirsh/678/base 2025-12-04T09:16:47.2284541Z * [new branch] gh/bdhirsh/678/head -> origin/gh/bdhirsh/678/head 2025-12-04T09:16:47.2286364Z * [new branch] gh/bdhirsh/678/orig -> origin/gh/bdhirsh/678/orig 2025-12-04T09:16:47.2289447Z * [new branch] gh/bdhirsh/679/base -> origin/gh/bdhirsh/679/base 2025-12-04T09:16:47.2291166Z * [new branch] gh/bdhirsh/679/head -> origin/gh/bdhirsh/679/head 2025-12-04T09:16:47.2292976Z * [new branch] gh/bdhirsh/679/orig -> origin/gh/bdhirsh/679/orig 2025-12-04T09:16:47.2296184Z * [new branch] gh/bdhirsh/680/base -> origin/gh/bdhirsh/680/base 2025-12-04T09:16:47.2298112Z * [new branch] gh/bdhirsh/680/head -> origin/gh/bdhirsh/680/head 2025-12-04T09:16:47.2299930Z * [new branch] gh/bdhirsh/680/orig -> origin/gh/bdhirsh/680/orig 2025-12-04T09:16:47.2302317Z * [new branch] gh/bdhirsh/681/base -> origin/gh/bdhirsh/681/base 2025-12-04T09:16:47.2304271Z * [new branch] gh/bdhirsh/681/head -> origin/gh/bdhirsh/681/head 2025-12-04T09:16:47.2306330Z * [new branch] gh/bdhirsh/681/orig -> origin/gh/bdhirsh/681/orig 2025-12-04T09:16:47.2309291Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-12-04T09:16:47.2311152Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-12-04T09:16:47.2313002Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-12-04T09:16:47.2315595Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-12-04T09:16:47.2317435Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-12-04T09:16:47.2319261Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-12-04T09:16:47.2321934Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-12-04T09:16:47.2323740Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-12-04T09:16:47.2325555Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-12-04T09:16:47.2328170Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-12-04T09:16:47.2330003Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-12-04T09:16:47.2331872Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-12-04T09:16:47.2334419Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-12-04T09:16:47.2337906Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-12-04T09:16:47.2338485Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-12-04T09:16:47.2341092Z * [new branch] gh/benjaminglass1/109/base -> origin/gh/benjaminglass1/109/base 2025-12-04T09:16:47.2342144Z * [new branch] gh/benjaminglass1/109/head -> origin/gh/benjaminglass1/109/head 2025-12-04T09:16:47.2344141Z * [new branch] gh/benjaminglass1/109/orig -> origin/gh/benjaminglass1/109/orig 2025-12-04T09:16:47.2346819Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-12-04T09:16:47.2348546Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-12-04T09:16:47.2350376Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-12-04T09:16:47.2353411Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-12-04T09:16:47.2355218Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-12-04T09:16:47.2357058Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-12-04T09:16:47.2359590Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-12-04T09:16:47.2361674Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-12-04T09:16:47.2363493Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-12-04T09:16:47.2365891Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-12-04T09:16:47.2367738Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-12-04T09:16:47.2369820Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-12-04T09:16:47.2372318Z * [new branch] gh/bobrenjc93/653/base -> origin/gh/bobrenjc93/653/base 2025-12-04T09:16:47.2374148Z * [new branch] gh/bobrenjc93/653/head -> origin/gh/bobrenjc93/653/head 2025-12-04T09:16:47.2375988Z * [new branch] gh/bobrenjc93/653/orig -> origin/gh/bobrenjc93/653/orig 2025-12-04T09:16:47.2379103Z * [new branch] gh/bobrenjc93/654/base -> origin/gh/bobrenjc93/654/base 2025-12-04T09:16:47.2380871Z * [new branch] gh/bobrenjc93/654/head -> origin/gh/bobrenjc93/654/head 2025-12-04T09:16:47.2383347Z * [new branch] gh/bobrenjc93/654/orig -> origin/gh/bobrenjc93/654/orig 2025-12-04T09:16:47.2386985Z * [new branch] gh/bobrenjc93/657/base -> origin/gh/bobrenjc93/657/base 2025-12-04T09:16:47.2389567Z * [new branch] gh/bobrenjc93/657/head -> origin/gh/bobrenjc93/657/head 2025-12-04T09:16:47.2392161Z * [new branch] gh/bobrenjc93/657/orig -> origin/gh/bobrenjc93/657/orig 2025-12-04T09:16:47.2395629Z * [new branch] gh/bobrenjc93/672/base -> origin/gh/bobrenjc93/672/base 2025-12-04T09:16:47.2397961Z * [new branch] gh/bobrenjc93/672/head -> origin/gh/bobrenjc93/672/head 2025-12-04T09:16:47.2400592Z * [new branch] gh/bobrenjc93/672/orig -> origin/gh/bobrenjc93/672/orig 2025-12-04T09:16:47.2404040Z * [new branch] gh/bobrenjc93/679/base -> origin/gh/bobrenjc93/679/base 2025-12-04T09:16:47.2406769Z * [new branch] gh/bobrenjc93/679/head -> origin/gh/bobrenjc93/679/head 2025-12-04T09:16:47.2409446Z * [new branch] gh/bobrenjc93/679/orig -> origin/gh/bobrenjc93/679/orig 2025-12-04T09:16:47.2412816Z * [new branch] gh/bobrenjc93/680/base -> origin/gh/bobrenjc93/680/base 2025-12-04T09:16:47.2415488Z * [new branch] gh/bobrenjc93/680/head -> origin/gh/bobrenjc93/680/head 2025-12-04T09:16:47.2417945Z * [new branch] gh/bobrenjc93/680/orig -> origin/gh/bobrenjc93/680/orig 2025-12-04T09:16:47.2421058Z * [new branch] gh/bobrenjc93/681/base -> origin/gh/bobrenjc93/681/base 2025-12-04T09:16:47.2423514Z * [new branch] gh/bobrenjc93/681/head -> origin/gh/bobrenjc93/681/head 2025-12-04T09:16:47.2425989Z * [new branch] gh/bobrenjc93/681/orig -> origin/gh/bobrenjc93/681/orig 2025-12-04T09:16:47.2429131Z * [new branch] gh/bobrenjc93/682/base -> origin/gh/bobrenjc93/682/base 2025-12-04T09:16:47.2431562Z * [new branch] gh/bobrenjc93/682/head -> origin/gh/bobrenjc93/682/head 2025-12-04T09:16:47.2434055Z * [new branch] gh/bobrenjc93/682/orig -> origin/gh/bobrenjc93/682/orig 2025-12-04T09:16:47.2437660Z * [new branch] gh/bobrenjc93/683/base -> origin/gh/bobrenjc93/683/base 2025-12-04T09:16:47.2439121Z * [new branch] gh/bobrenjc93/683/head -> origin/gh/bobrenjc93/683/head 2025-12-04T09:16:47.2440954Z * [new branch] gh/bobrenjc93/683/orig -> origin/gh/bobrenjc93/683/orig 2025-12-04T09:16:47.2443505Z * [new branch] gh/bobrenjc93/684/base -> origin/gh/bobrenjc93/684/base 2025-12-04T09:16:47.2445513Z * [new branch] gh/bobrenjc93/684/head -> origin/gh/bobrenjc93/684/head 2025-12-04T09:16:47.2447597Z * [new branch] gh/bobrenjc93/684/orig -> origin/gh/bobrenjc93/684/orig 2025-12-04T09:16:47.2450165Z * [new branch] gh/bobrenjc93/685/base -> origin/gh/bobrenjc93/685/base 2025-12-04T09:16:47.2452236Z * [new branch] gh/bobrenjc93/685/head -> origin/gh/bobrenjc93/685/head 2025-12-04T09:16:47.2454416Z * [new branch] gh/bobrenjc93/685/orig -> origin/gh/bobrenjc93/685/orig 2025-12-04T09:16:47.2457205Z * [new branch] gh/bobrenjc93/686/base -> origin/gh/bobrenjc93/686/base 2025-12-04T09:16:47.2460406Z * [new branch] gh/bobrenjc93/686/head -> origin/gh/bobrenjc93/686/head 2025-12-04T09:16:47.2461711Z * [new branch] gh/bobrenjc93/686/orig -> origin/gh/bobrenjc93/686/orig 2025-12-04T09:16:47.2464135Z * [new branch] gh/bobrenjc93/687/base -> origin/gh/bobrenjc93/687/base 2025-12-04T09:16:47.2466392Z * [new branch] gh/bobrenjc93/687/head -> origin/gh/bobrenjc93/687/head 2025-12-04T09:16:47.2468137Z * [new branch] gh/bobrenjc93/687/orig -> origin/gh/bobrenjc93/687/orig 2025-12-04T09:16:47.2471434Z * [new branch] gh/bobrenjc93/688/base -> origin/gh/bobrenjc93/688/base 2025-12-04T09:16:47.2473282Z * [new branch] gh/bobrenjc93/688/head -> origin/gh/bobrenjc93/688/head 2025-12-04T09:16:47.2475140Z * [new branch] gh/bobrenjc93/688/orig -> origin/gh/bobrenjc93/688/orig 2025-12-04T09:16:47.2477578Z * [new branch] gh/bobrenjc93/689/base -> origin/gh/bobrenjc93/689/base 2025-12-04T09:16:47.2479519Z * [new branch] gh/bobrenjc93/689/head -> origin/gh/bobrenjc93/689/head 2025-12-04T09:16:47.2481382Z * [new branch] gh/bobrenjc93/689/orig -> origin/gh/bobrenjc93/689/orig 2025-12-04T09:16:47.2483789Z * [new branch] gh/bobrenjc93/690/base -> origin/gh/bobrenjc93/690/base 2025-12-04T09:16:47.2485628Z * [new branch] gh/bobrenjc93/690/head -> origin/gh/bobrenjc93/690/head 2025-12-04T09:16:47.2487643Z * [new branch] gh/bobrenjc93/690/orig -> origin/gh/bobrenjc93/690/orig 2025-12-04T09:16:47.2491035Z * [new branch] gh/bobrenjc93/691/base -> origin/gh/bobrenjc93/691/base 2025-12-04T09:16:47.2493053Z * [new branch] gh/bobrenjc93/691/head -> origin/gh/bobrenjc93/691/head 2025-12-04T09:16:47.2495295Z * [new branch] gh/bobrenjc93/691/orig -> origin/gh/bobrenjc93/691/orig 2025-12-04T09:16:47.2498543Z * [new branch] gh/bobrenjc93/692/base -> origin/gh/bobrenjc93/692/base 2025-12-04T09:16:47.2500378Z * [new branch] gh/bobrenjc93/692/head -> origin/gh/bobrenjc93/692/head 2025-12-04T09:16:47.2502219Z * [new branch] gh/bobrenjc93/692/orig -> origin/gh/bobrenjc93/692/orig 2025-12-04T09:16:47.2504659Z * [new branch] gh/bobrenjc93/693/base -> origin/gh/bobrenjc93/693/base 2025-12-04T09:16:47.2506481Z * [new branch] gh/bobrenjc93/693/head -> origin/gh/bobrenjc93/693/head 2025-12-04T09:16:47.2508405Z * [new branch] gh/bobrenjc93/693/orig -> origin/gh/bobrenjc93/693/orig 2025-12-04T09:16:47.2511061Z * [new branch] gh/bobrenjc93/694/base -> origin/gh/bobrenjc93/694/base 2025-12-04T09:16:47.2512959Z * [new branch] gh/bobrenjc93/694/head -> origin/gh/bobrenjc93/694/head 2025-12-04T09:16:47.2514919Z * [new branch] gh/bobrenjc93/694/orig -> origin/gh/bobrenjc93/694/orig 2025-12-04T09:16:47.2517370Z * [new branch] gh/bobrenjc93/695/base -> origin/gh/bobrenjc93/695/base 2025-12-04T09:16:47.2519228Z * [new branch] gh/bobrenjc93/695/head -> origin/gh/bobrenjc93/695/head 2025-12-04T09:16:47.2521058Z * [new branch] gh/bobrenjc93/695/orig -> origin/gh/bobrenjc93/695/orig 2025-12-04T09:16:47.2524205Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-12-04T09:16:47.2526082Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-12-04T09:16:47.2528842Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-12-04T09:16:47.2530571Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-12-04T09:16:47.2532381Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-12-04T09:16:47.2534765Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-12-04T09:16:47.2537043Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-12-04T09:16:47.2539164Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-12-04T09:16:47.2541787Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-12-04T09:16:47.2543792Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-12-04T09:16:47.2545086Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-12-04T09:16:47.2548058Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-12-04T09:16:47.2549392Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-12-04T09:16:47.2551743Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-12-04T09:16:47.2554053Z * [new branch] gh/c00w/58/base -> origin/gh/c00w/58/base 2025-12-04T09:16:47.2555835Z * [new branch] gh/c00w/58/head -> origin/gh/c00w/58/head 2025-12-04T09:16:47.2557589Z * [new branch] gh/c00w/58/orig -> origin/gh/c00w/58/orig 2025-12-04T09:16:47.2560860Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-12-04T09:16:47.2562875Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-12-04T09:16:47.2564418Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-12-04T09:16:47.2567681Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-12-04T09:16:47.2569642Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-12-04T09:16:47.2572445Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-12-04T09:16:47.2574199Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-12-04T09:16:47.2576192Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-12-04T09:16:47.2578741Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-12-04T09:16:47.2580865Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-12-04T09:16:47.2582664Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-12-04T09:16:47.2585191Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-12-04T09:16:47.2587246Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-12-04T09:16:47.2589100Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-12-04T09:16:47.2591520Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-12-04T09:16:47.2593575Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-12-04T09:16:47.2595108Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-12-04T09:16:47.2597797Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-12-04T09:16:47.2599996Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-12-04T09:16:47.2602509Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-12-04T09:16:47.2605102Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-12-04T09:16:47.2607021Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-12-04T09:16:47.2609093Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-12-04T09:16:47.2612211Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-12-04T09:16:47.2614981Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-12-04T09:16:47.2617517Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-12-04T09:16:47.2620921Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-12-04T09:16:47.2623708Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-12-04T09:16:47.2626008Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-12-04T09:16:47.2629485Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-12-04T09:16:47.2631996Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-12-04T09:16:47.2634701Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-12-04T09:16:47.2639561Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-12-04T09:16:47.2641182Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-12-04T09:16:47.2643039Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-12-04T09:16:47.2645741Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-12-04T09:16:47.2647708Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-12-04T09:16:47.2649668Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-12-04T09:16:47.2652288Z * [new branch] gh/coconutruben/84/base -> origin/gh/coconutruben/84/base 2025-12-04T09:16:47.2654243Z * [new branch] gh/coconutruben/84/head -> origin/gh/coconutruben/84/head 2025-12-04T09:16:47.2656051Z * [new branch] gh/coconutruben/84/orig -> origin/gh/coconutruben/84/orig 2025-12-04T09:16:47.2658641Z * [new branch] gh/coconutruben/85/base -> origin/gh/coconutruben/85/base 2025-12-04T09:16:47.2660538Z * [new branch] gh/coconutruben/85/head -> origin/gh/coconutruben/85/head 2025-12-04T09:16:47.2662395Z * [new branch] gh/coconutruben/85/orig -> origin/gh/coconutruben/85/orig 2025-12-04T09:16:47.2665070Z * [new branch] gh/coconutruben/86/base -> origin/gh/coconutruben/86/base 2025-12-04T09:16:47.2667156Z * [new branch] gh/coconutruben/86/head -> origin/gh/coconutruben/86/head 2025-12-04T09:16:47.2668689Z * [new branch] gh/coconutruben/86/orig -> origin/gh/coconutruben/86/orig 2025-12-04T09:16:47.2671947Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-12-04T09:16:47.2673813Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-12-04T09:16:47.2676155Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-12-04T09:16:47.2677928Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-12-04T09:16:47.2680499Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-12-04T09:16:47.2682322Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-12-04T09:16:47.2684715Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-12-04T09:16:47.2686541Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-12-04T09:16:47.2689815Z * [new branch] gh/d4l3k/1/base -> origin/gh/d4l3k/1/base 2025-12-04T09:16:47.2691621Z * [new branch] gh/d4l3k/1/head -> origin/gh/d4l3k/1/head 2025-12-04T09:16:47.2694248Z * [new branch] gh/d4l3k/2/base -> origin/gh/d4l3k/2/base 2025-12-04T09:16:47.2696061Z * [new branch] gh/d4l3k/2/head -> origin/gh/d4l3k/2/head 2025-12-04T09:16:47.2697837Z * [new branch] gh/d4l3k/2/orig -> origin/gh/d4l3k/2/orig 2025-12-04T09:16:47.2700412Z * [new branch] gh/d4l3k/3/base -> origin/gh/d4l3k/3/base 2025-12-04T09:16:47.2702133Z * [new branch] gh/d4l3k/3/head -> origin/gh/d4l3k/3/head 2025-12-04T09:16:47.2704166Z * [new branch] gh/d4l3k/3/orig -> origin/gh/d4l3k/3/orig 2025-12-04T09:16:47.2706546Z * [new branch] gh/d4l3k/4/base -> origin/gh/d4l3k/4/base 2025-12-04T09:16:47.2708450Z * [new branch] gh/d4l3k/4/head -> origin/gh/d4l3k/4/head 2025-12-04T09:16:47.2710303Z * [new branch] gh/d4l3k/4/orig -> origin/gh/d4l3k/4/orig 2025-12-04T09:16:47.2712716Z * [new branch] gh/d4l3k/5/base -> origin/gh/d4l3k/5/base 2025-12-04T09:16:47.2714535Z * [new branch] gh/d4l3k/5/orig -> origin/gh/d4l3k/5/orig 2025-12-04T09:16:47.2717829Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-12-04T09:16:47.2719726Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-12-04T09:16:47.2721623Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-12-04T09:16:47.2724164Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-12-04T09:16:47.2726065Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-12-04T09:16:47.2727958Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-12-04T09:16:47.2731039Z * [new branch] gh/desertfire/605/base -> origin/gh/desertfire/605/base 2025-12-04T09:16:47.2732837Z * [new branch] gh/desertfire/605/head -> origin/gh/desertfire/605/head 2025-12-04T09:16:47.2734740Z * [new branch] gh/desertfire/605/orig -> origin/gh/desertfire/605/orig 2025-12-04T09:16:47.2737566Z * [new branch] gh/desertfire/606/base -> origin/gh/desertfire/606/base 2025-12-04T09:16:47.2739228Z * [new branch] gh/desertfire/606/head -> origin/gh/desertfire/606/head 2025-12-04T09:16:47.2741252Z * [new branch] gh/desertfire/606/orig -> origin/gh/desertfire/606/orig 2025-12-04T09:16:47.2743789Z * [new branch] gh/desertfire/607/base -> origin/gh/desertfire/607/base 2025-12-04T09:16:47.2745615Z * [new branch] gh/desertfire/607/head -> origin/gh/desertfire/607/head 2025-12-04T09:16:47.2747530Z * [new branch] gh/desertfire/607/orig -> origin/gh/desertfire/607/orig 2025-12-04T09:16:47.2750078Z * [new branch] gh/desertfire/608/base -> origin/gh/desertfire/608/base 2025-12-04T09:16:47.2751897Z * [new branch] gh/desertfire/608/head -> origin/gh/desertfire/608/head 2025-12-04T09:16:47.2753743Z * [new branch] gh/desertfire/608/orig -> origin/gh/desertfire/608/orig 2025-12-04T09:16:47.2756228Z * [new branch] gh/desertfire/609/base -> origin/gh/desertfire/609/base 2025-12-04T09:16:47.2758092Z * [new branch] gh/desertfire/609/head -> origin/gh/desertfire/609/head 2025-12-04T09:16:47.2760009Z * [new branch] gh/desertfire/609/orig -> origin/gh/desertfire/609/orig 2025-12-04T09:16:47.2763027Z * [new branch] gh/desertfire/610/base -> origin/gh/desertfire/610/base 2025-12-04T09:16:47.2764890Z * [new branch] gh/desertfire/610/head -> origin/gh/desertfire/610/head 2025-12-04T09:16:47.2766751Z * [new branch] gh/desertfire/610/orig -> origin/gh/desertfire/610/orig 2025-12-04T09:16:47.2769481Z * [new branch] gh/desertfire/611/base -> origin/gh/desertfire/611/base 2025-12-04T09:16:47.2771381Z * [new branch] gh/desertfire/611/head -> origin/gh/desertfire/611/head 2025-12-04T09:16:47.2773208Z * [new branch] gh/desertfire/611/orig -> origin/gh/desertfire/611/orig 2025-12-04T09:16:47.2775772Z * [new branch] gh/desertfire/612/base -> origin/gh/desertfire/612/base 2025-12-04T09:16:47.2777788Z * [new branch] gh/desertfire/612/head -> origin/gh/desertfire/612/head 2025-12-04T09:16:47.2779509Z * [new branch] gh/desertfire/612/orig -> origin/gh/desertfire/612/orig 2025-12-04T09:16:47.2782073Z * [new branch] gh/desertfire/613/base -> origin/gh/desertfire/613/base 2025-12-04T09:16:47.2783986Z * [new branch] gh/desertfire/613/head -> origin/gh/desertfire/613/head 2025-12-04T09:16:47.2785861Z * [new branch] gh/desertfire/613/orig -> origin/gh/desertfire/613/orig 2025-12-04T09:16:47.2788521Z * [new branch] gh/desertfire/614/base -> origin/gh/desertfire/614/base 2025-12-04T09:16:47.2790462Z * [new branch] gh/desertfire/614/head -> origin/gh/desertfire/614/head 2025-12-04T09:16:47.2792291Z * [new branch] gh/desertfire/614/orig -> origin/gh/desertfire/614/orig 2025-12-04T09:16:47.2794920Z * [new branch] gh/desertfire/615/base -> origin/gh/desertfire/615/base 2025-12-04T09:16:47.2797021Z * [new branch] gh/desertfire/615/head -> origin/gh/desertfire/615/head 2025-12-04T09:16:47.2798807Z * [new branch] gh/desertfire/615/orig -> origin/gh/desertfire/615/orig 2025-12-04T09:16:47.2801203Z * [new branch] gh/desertfire/616/base -> origin/gh/desertfire/616/base 2025-12-04T09:16:47.2803141Z * [new branch] gh/desertfire/616/head -> origin/gh/desertfire/616/head 2025-12-04T09:16:47.2804905Z * [new branch] gh/desertfire/616/orig -> origin/gh/desertfire/616/orig 2025-12-04T09:16:47.2807277Z * [new branch] gh/desertfire/617/base -> origin/gh/desertfire/617/base 2025-12-04T09:16:47.2809364Z * [new branch] gh/desertfire/617/head -> origin/gh/desertfire/617/head 2025-12-04T09:16:47.2811103Z * [new branch] gh/desertfire/617/orig -> origin/gh/desertfire/617/orig 2025-12-04T09:16:47.2814241Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-12-04T09:16:47.2816188Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-12-04T09:16:47.2819326Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-12-04T09:16:47.2821200Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-12-04T09:16:47.2823019Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-12-04T09:16:47.2825531Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-12-04T09:16:47.2827351Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-12-04T09:16:47.2829722Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-12-04T09:16:47.2831506Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-12-04T09:16:47.2833833Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-12-04T09:16:47.2835848Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-12-04T09:16:47.2838576Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-12-04T09:16:47.2840352Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-12-04T09:16:47.2842817Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-12-04T09:16:47.2844656Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-12-04T09:16:47.2846476Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-12-04T09:16:47.2849444Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-12-04T09:16:47.2850951Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-12-04T09:16:47.2852864Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-12-04T09:16:47.2855221Z * [new branch] gh/drisspg/218/base -> origin/gh/drisspg/218/base 2025-12-04T09:16:47.2857046Z * [new branch] gh/drisspg/218/head -> origin/gh/drisspg/218/head 2025-12-04T09:16:47.2858884Z * [new branch] gh/drisspg/218/orig -> origin/gh/drisspg/218/orig 2025-12-04T09:16:47.2861383Z * [new branch] gh/drisspg/219/base -> origin/gh/drisspg/219/base 2025-12-04T09:16:47.2863232Z * [new branch] gh/drisspg/219/head -> origin/gh/drisspg/219/head 2025-12-04T09:16:47.2865135Z * [new branch] gh/drisspg/219/orig -> origin/gh/drisspg/219/orig 2025-12-04T09:16:47.2867636Z * [new branch] gh/drisspg/220/base -> origin/gh/drisspg/220/base 2025-12-04T09:16:47.2869424Z * [new branch] gh/drisspg/220/head -> origin/gh/drisspg/220/head 2025-12-04T09:16:47.2871261Z * [new branch] gh/drisspg/220/orig -> origin/gh/drisspg/220/orig 2025-12-04T09:16:47.2873792Z * [new branch] gh/drisspg/221/base -> origin/gh/drisspg/221/base 2025-12-04T09:16:47.2875619Z * [new branch] gh/drisspg/221/head -> origin/gh/drisspg/221/head 2025-12-04T09:16:47.2877478Z * [new branch] gh/drisspg/221/orig -> origin/gh/drisspg/221/orig 2025-12-04T09:16:47.2880014Z * [new branch] gh/drisspg/222/base -> origin/gh/drisspg/222/base 2025-12-04T09:16:47.2881833Z * [new branch] gh/drisspg/222/head -> origin/gh/drisspg/222/head 2025-12-04T09:16:47.2883655Z * [new branch] gh/drisspg/222/orig -> origin/gh/drisspg/222/orig 2025-12-04T09:16:47.2886206Z * [new branch] gh/drisspg/223/base -> origin/gh/drisspg/223/base 2025-12-04T09:16:47.2888165Z * [new branch] gh/drisspg/223/head -> origin/gh/drisspg/223/head 2025-12-04T09:16:47.2890070Z * [new branch] gh/drisspg/223/orig -> origin/gh/drisspg/223/orig 2025-12-04T09:16:47.2892519Z * [new branch] gh/drisspg/224/base -> origin/gh/drisspg/224/base 2025-12-04T09:16:47.2894335Z * [new branch] gh/drisspg/224/head -> origin/gh/drisspg/224/head 2025-12-04T09:16:47.2896217Z * [new branch] gh/drisspg/224/orig -> origin/gh/drisspg/224/orig 2025-12-04T09:16:47.2898704Z * [new branch] gh/drisspg/225/base -> origin/gh/drisspg/225/base 2025-12-04T09:16:47.2909334Z * [new branch] gh/drisspg/225/head -> origin/gh/drisspg/225/head 2025-12-04T09:16:47.2909822Z * [new branch] gh/drisspg/225/orig -> origin/gh/drisspg/225/orig 2025-12-04T09:16:47.2910053Z * [new branch] gh/drisspg/226/base -> origin/gh/drisspg/226/base 2025-12-04T09:16:47.2910270Z * [new branch] gh/drisspg/226/head -> origin/gh/drisspg/226/head 2025-12-04T09:16:47.2910492Z * [new branch] gh/drisspg/226/orig -> origin/gh/drisspg/226/orig 2025-12-04T09:16:47.2911446Z * [new branch] gh/drisspg/227/base -> origin/gh/drisspg/227/base 2025-12-04T09:16:47.2913420Z * [new branch] gh/drisspg/227/head -> origin/gh/drisspg/227/head 2025-12-04T09:16:47.2915323Z * [new branch] gh/drisspg/227/orig -> origin/gh/drisspg/227/orig 2025-12-04T09:16:47.2917896Z * [new branch] gh/drisspg/228/base -> origin/gh/drisspg/228/base 2025-12-04T09:16:47.2919744Z * [new branch] gh/drisspg/228/head -> origin/gh/drisspg/228/head 2025-12-04T09:16:47.2921555Z * [new branch] gh/drisspg/228/orig -> origin/gh/drisspg/228/orig 2025-12-04T09:16:47.2924091Z * [new branch] gh/drisspg/229/base -> origin/gh/drisspg/229/base 2025-12-04T09:16:47.2925937Z * [new branch] gh/drisspg/229/head -> origin/gh/drisspg/229/head 2025-12-04T09:16:47.2928026Z * [new branch] gh/drisspg/229/orig -> origin/gh/drisspg/229/orig 2025-12-04T09:16:47.2930690Z * [new branch] gh/drisspg/230/base -> origin/gh/drisspg/230/base 2025-12-04T09:16:47.2932899Z * [new branch] gh/drisspg/230/head -> origin/gh/drisspg/230/head 2025-12-04T09:16:47.2934469Z * [new branch] gh/drisspg/230/orig -> origin/gh/drisspg/230/orig 2025-12-04T09:16:47.2937936Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-12-04T09:16:47.2940212Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-12-04T09:16:47.2943056Z * [new branch] gh/dzmitry-huba/1/base -> origin/gh/dzmitry-huba/1/base 2025-12-04T09:16:47.2945024Z * [new branch] gh/dzmitry-huba/1/head -> origin/gh/dzmitry-huba/1/head 2025-12-04T09:16:47.2947651Z * [new branch] gh/dzmitry-huba/12/base -> origin/gh/dzmitry-huba/12/base 2025-12-04T09:16:47.2949499Z * [new branch] gh/dzmitry-huba/12/head -> origin/gh/dzmitry-huba/12/head 2025-12-04T09:16:47.2951360Z * [new branch] gh/dzmitry-huba/12/orig -> origin/gh/dzmitry-huba/12/orig 2025-12-04T09:16:47.2953981Z * [new branch] gh/dzmitry-huba/13/base -> origin/gh/dzmitry-huba/13/base 2025-12-04T09:16:47.2955856Z * [new branch] gh/dzmitry-huba/13/head -> origin/gh/dzmitry-huba/13/head 2025-12-04T09:16:47.2957783Z * [new branch] gh/dzmitry-huba/13/orig -> origin/gh/dzmitry-huba/13/orig 2025-12-04T09:16:47.2960103Z * [new branch] gh/dzmitry-huba/14/base -> origin/gh/dzmitry-huba/14/base 2025-12-04T09:16:47.2961951Z * [new branch] gh/dzmitry-huba/14/head -> origin/gh/dzmitry-huba/14/head 2025-12-04T09:16:47.2963945Z * [new branch] gh/dzmitry-huba/14/orig -> origin/gh/dzmitry-huba/14/orig 2025-12-04T09:16:47.2966500Z * [new branch] gh/dzmitry-huba/15/base -> origin/gh/dzmitry-huba/15/base 2025-12-04T09:16:47.2968604Z * [new branch] gh/dzmitry-huba/15/head -> origin/gh/dzmitry-huba/15/head 2025-12-04T09:16:47.2970296Z * [new branch] gh/dzmitry-huba/15/orig -> origin/gh/dzmitry-huba/15/orig 2025-12-04T09:16:47.2973275Z * [new branch] gh/dzmitry-huba/16/base -> origin/gh/dzmitry-huba/16/base 2025-12-04T09:16:47.2974843Z * [new branch] gh/dzmitry-huba/16/head -> origin/gh/dzmitry-huba/16/head 2025-12-04T09:16:47.2976613Z * [new branch] gh/dzmitry-huba/16/orig -> origin/gh/dzmitry-huba/16/orig 2025-12-04T09:16:47.2979217Z * [new branch] gh/dzmitry-huba/17/base -> origin/gh/dzmitry-huba/17/base 2025-12-04T09:16:47.2981099Z * [new branch] gh/dzmitry-huba/17/head -> origin/gh/dzmitry-huba/17/head 2025-12-04T09:16:47.2982935Z * [new branch] gh/dzmitry-huba/17/orig -> origin/gh/dzmitry-huba/17/orig 2025-12-04T09:16:47.2985279Z * [new branch] gh/dzmitry-huba/2/base -> origin/gh/dzmitry-huba/2/base 2025-12-04T09:16:47.2987066Z * [new branch] gh/dzmitry-huba/2/head -> origin/gh/dzmitry-huba/2/head 2025-12-04T09:16:47.2989570Z * [new branch] gh/dzmitry-huba/3/base -> origin/gh/dzmitry-huba/3/base 2025-12-04T09:16:47.2991377Z * [new branch] gh/dzmitry-huba/3/head -> origin/gh/dzmitry-huba/3/head 2025-12-04T09:16:47.2994478Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-12-04T09:16:47.2996419Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-12-04T09:16:47.2998219Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-12-04T09:16:47.3001036Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-12-04T09:16:47.3003010Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-12-04T09:16:47.3004691Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-12-04T09:16:47.3007197Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-12-04T09:16:47.3009219Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-12-04T09:16:47.3011000Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-12-04T09:16:47.3013578Z * [new branch] gh/eellison/862/base -> origin/gh/eellison/862/base 2025-12-04T09:16:47.3015467Z * [new branch] gh/eellison/862/head -> origin/gh/eellison/862/head 2025-12-04T09:16:47.3017302Z * [new branch] gh/eellison/862/orig -> origin/gh/eellison/862/orig 2025-12-04T09:16:47.3019816Z * [new branch] gh/eellison/863/base -> origin/gh/eellison/863/base 2025-12-04T09:16:47.3021658Z * [new branch] gh/eellison/863/head -> origin/gh/eellison/863/head 2025-12-04T09:16:47.3023517Z * [new branch] gh/eellison/863/orig -> origin/gh/eellison/863/orig 2025-12-04T09:16:47.3025916Z * [new branch] gh/eellison/864/base -> origin/gh/eellison/864/base 2025-12-04T09:16:47.3027794Z * [new branch] gh/eellison/864/head -> origin/gh/eellison/864/head 2025-12-04T09:16:47.3030209Z * [new branch] gh/eellison/864/orig -> origin/gh/eellison/864/orig 2025-12-04T09:16:47.3032799Z * [new branch] gh/eellison/865/base -> origin/gh/eellison/865/base 2025-12-04T09:16:47.3034565Z * [new branch] gh/eellison/865/head -> origin/gh/eellison/865/head 2025-12-04T09:16:47.3036696Z * [new branch] gh/eellison/865/orig -> origin/gh/eellison/865/orig 2025-12-04T09:16:47.3041940Z * [new branch] gh/eellison/866/base -> origin/gh/eellison/866/base 2025-12-04T09:16:47.3043645Z * [new branch] gh/eellison/866/head -> origin/gh/eellison/866/head 2025-12-04T09:16:47.3045488Z * [new branch] gh/eellison/866/orig -> origin/gh/eellison/866/orig 2025-12-04T09:16:47.3048283Z * [new branch] gh/eellison/867/base -> origin/gh/eellison/867/base 2025-12-04T09:16:47.3050040Z * [new branch] gh/eellison/867/head -> origin/gh/eellison/867/head 2025-12-04T09:16:47.3051972Z * [new branch] gh/eellison/867/orig -> origin/gh/eellison/867/orig 2025-12-04T09:16:47.3054677Z * [new branch] gh/eellison/868/base -> origin/gh/eellison/868/base 2025-12-04T09:16:47.3056801Z * [new branch] gh/eellison/868/head -> origin/gh/eellison/868/head 2025-12-04T09:16:47.3058524Z * [new branch] gh/eellison/868/orig -> origin/gh/eellison/868/orig 2025-12-04T09:16:47.3061194Z * [new branch] gh/eellison/869/base -> origin/gh/eellison/869/base 2025-12-04T09:16:47.3062961Z * [new branch] gh/eellison/869/head -> origin/gh/eellison/869/head 2025-12-04T09:16:47.3064754Z * [new branch] gh/eellison/869/orig -> origin/gh/eellison/869/orig 2025-12-04T09:16:47.3067416Z * [new branch] gh/eellison/870/base -> origin/gh/eellison/870/base 2025-12-04T09:16:47.3069216Z * [new branch] gh/eellison/870/head -> origin/gh/eellison/870/head 2025-12-04T09:16:47.3071017Z * [new branch] gh/eellison/870/orig -> origin/gh/eellison/870/orig 2025-12-04T09:16:47.3073562Z * [new branch] gh/eellison/871/base -> origin/gh/eellison/871/base 2025-12-04T09:16:47.3075348Z * [new branch] gh/eellison/871/head -> origin/gh/eellison/871/head 2025-12-04T09:16:47.3077612Z * [new branch] gh/eellison/871/orig -> origin/gh/eellison/871/orig 2025-12-04T09:16:47.3079817Z * [new branch] gh/eellison/872/base -> origin/gh/eellison/872/base 2025-12-04T09:16:47.3081483Z * [new branch] gh/eellison/872/head -> origin/gh/eellison/872/head 2025-12-04T09:16:47.3083379Z * [new branch] gh/eellison/872/orig -> origin/gh/eellison/872/orig 2025-12-04T09:16:47.3086072Z * [new branch] gh/eellison/873/base -> origin/gh/eellison/873/base 2025-12-04T09:16:47.3087877Z * [new branch] gh/eellison/873/head -> origin/gh/eellison/873/head 2025-12-04T09:16:47.3089787Z * [new branch] gh/eellison/873/orig -> origin/gh/eellison/873/orig 2025-12-04T09:16:47.3092824Z * [new branch] gh/eellison/874/base -> origin/gh/eellison/874/base 2025-12-04T09:16:47.3094717Z * [new branch] gh/eellison/874/head -> origin/gh/eellison/874/head 2025-12-04T09:16:47.3096570Z * [new branch] gh/eellison/874/orig -> origin/gh/eellison/874/orig 2025-12-04T09:16:47.3099762Z * [new branch] gh/eellison/875/base -> origin/gh/eellison/875/base 2025-12-04T09:16:47.3101622Z * [new branch] gh/eellison/875/head -> origin/gh/eellison/875/head 2025-12-04T09:16:47.3103425Z * [new branch] gh/eellison/875/orig -> origin/gh/eellison/875/orig 2025-12-04T09:16:47.3106059Z * [new branch] gh/eellison/876/base -> origin/gh/eellison/876/base 2025-12-04T09:16:47.3107981Z * [new branch] gh/eellison/876/head -> origin/gh/eellison/876/head 2025-12-04T09:16:47.3109845Z * [new branch] gh/eellison/876/orig -> origin/gh/eellison/876/orig 2025-12-04T09:16:47.3112403Z * [new branch] gh/eellison/877/base -> origin/gh/eellison/877/base 2025-12-04T09:16:47.3114223Z * [new branch] gh/eellison/877/head -> origin/gh/eellison/877/head 2025-12-04T09:16:47.3116045Z * [new branch] gh/eellison/877/orig -> origin/gh/eellison/877/orig 2025-12-04T09:16:47.3118893Z * [new branch] gh/eellison/878/base -> origin/gh/eellison/878/base 2025-12-04T09:16:47.3120602Z * [new branch] gh/eellison/878/head -> origin/gh/eellison/878/head 2025-12-04T09:16:47.3122407Z * [new branch] gh/eellison/878/orig -> origin/gh/eellison/878/orig 2025-12-04T09:16:47.3125127Z * [new branch] gh/eellison/879/base -> origin/gh/eellison/879/base 2025-12-04T09:16:47.3126896Z * [new branch] gh/eellison/879/head -> origin/gh/eellison/879/head 2025-12-04T09:16:47.3128876Z * [new branch] gh/eellison/879/orig -> origin/gh/eellison/879/orig 2025-12-04T09:16:47.3131347Z * [new branch] gh/eellison/880/base -> origin/gh/eellison/880/base 2025-12-04T09:16:47.3133193Z * [new branch] gh/eellison/880/head -> origin/gh/eellison/880/head 2025-12-04T09:16:47.3135078Z * [new branch] gh/eellison/880/orig -> origin/gh/eellison/880/orig 2025-12-04T09:16:47.3138104Z * [new branch] gh/eellison/881/base -> origin/gh/eellison/881/base 2025-12-04T09:16:47.3139921Z * [new branch] gh/eellison/881/head -> origin/gh/eellison/881/head 2025-12-04T09:16:47.3141679Z * [new branch] gh/eellison/881/orig -> origin/gh/eellison/881/orig 2025-12-04T09:16:47.3144299Z * [new branch] gh/eellison/882/base -> origin/gh/eellison/882/base 2025-12-04T09:16:47.3146193Z * [new branch] gh/eellison/882/head -> origin/gh/eellison/882/head 2025-12-04T09:16:47.3148208Z * [new branch] gh/eellison/882/orig -> origin/gh/eellison/882/orig 2025-12-04T09:16:47.3150785Z * [new branch] gh/eellison/883/base -> origin/gh/eellison/883/base 2025-12-04T09:16:47.3152662Z * [new branch] gh/eellison/883/head -> origin/gh/eellison/883/head 2025-12-04T09:16:47.3154598Z * [new branch] gh/eellison/883/orig -> origin/gh/eellison/883/orig 2025-12-04T09:16:47.3156922Z * [new branch] gh/eellison/884/base -> origin/gh/eellison/884/base 2025-12-04T09:16:47.3158744Z * [new branch] gh/eellison/884/head -> origin/gh/eellison/884/head 2025-12-04T09:16:47.3160491Z * [new branch] gh/eellison/884/orig -> origin/gh/eellison/884/orig 2025-12-04T09:16:47.3163556Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-12-04T09:16:47.3165499Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-12-04T09:16:47.3168413Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-12-04T09:16:47.3170797Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-12-04T09:16:47.3172399Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-12-04T09:16:47.3174840Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-12-04T09:16:47.3176683Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-12-04T09:16:47.3178521Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-12-04T09:16:47.3181277Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-12-04T09:16:47.3183117Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-12-04T09:16:47.3184959Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-12-04T09:16:47.3187527Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-12-04T09:16:47.3189480Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-12-04T09:16:47.3191370Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-12-04T09:16:47.3193999Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-12-04T09:16:47.3195895Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-12-04T09:16:47.3197699Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-12-04T09:16:47.3200312Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-12-04T09:16:47.3202189Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-12-04T09:16:47.3204080Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-12-04T09:16:47.3206711Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-12-04T09:16:47.3208811Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-12-04T09:16:47.3210680Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-12-04T09:16:47.3213219Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-12-04T09:16:47.3215195Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-12-04T09:16:47.3217142Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-12-04T09:16:47.3219746Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-12-04T09:16:47.3221584Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-12-04T09:16:47.3223411Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-12-04T09:16:47.3225988Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-12-04T09:16:47.3227959Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-12-04T09:16:47.3229744Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-12-04T09:16:47.3232410Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-12-04T09:16:47.3234170Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-12-04T09:16:47.3236364Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-12-04T09:16:47.3239546Z * [new branch] gh/etaf/173/base -> origin/gh/etaf/173/base 2025-12-04T09:16:47.3241545Z * [new branch] gh/etaf/173/head -> origin/gh/etaf/173/head 2025-12-04T09:16:47.3243401Z * [new branch] gh/etaf/173/orig -> origin/gh/etaf/173/orig 2025-12-04T09:16:47.3246075Z * [new branch] gh/etaf/174/base -> origin/gh/etaf/174/base 2025-12-04T09:16:47.3248023Z * [new branch] gh/etaf/174/head -> origin/gh/etaf/174/head 2025-12-04T09:16:47.3250649Z * [new branch] gh/etaf/175/base -> origin/gh/etaf/175/base 2025-12-04T09:16:47.3252485Z * [new branch] gh/etaf/175/head -> origin/gh/etaf/175/head 2025-12-04T09:16:47.3254178Z * [new branch] gh/etaf/175/orig -> origin/gh/etaf/175/orig 2025-12-04T09:16:47.3256740Z * [new branch] gh/etaf/176/base -> origin/gh/etaf/176/base 2025-12-04T09:16:47.3258692Z * [new branch] gh/etaf/176/head -> origin/gh/etaf/176/head 2025-12-04T09:16:47.3260487Z * [new branch] gh/etaf/176/orig -> origin/gh/etaf/176/orig 2025-12-04T09:16:47.3263555Z * [new branch] gh/etaf/177/base -> origin/gh/etaf/177/base 2025-12-04T09:16:47.3265541Z * [new branch] gh/etaf/177/head -> origin/gh/etaf/177/head 2025-12-04T09:16:47.3267385Z * [new branch] gh/etaf/177/orig -> origin/gh/etaf/177/orig 2025-12-04T09:16:47.3270233Z * [new branch] gh/etaf/178/base -> origin/gh/etaf/178/base 2025-12-04T09:16:47.3272207Z * [new branch] gh/etaf/178/head -> origin/gh/etaf/178/head 2025-12-04T09:16:47.3274075Z * [new branch] gh/etaf/178/orig -> origin/gh/etaf/178/orig 2025-12-04T09:16:47.3276665Z * [new branch] gh/etaf/179/base -> origin/gh/etaf/179/base 2025-12-04T09:16:47.3278517Z * [new branch] gh/etaf/179/head -> origin/gh/etaf/179/head 2025-12-04T09:16:47.3280391Z * [new branch] gh/etaf/179/orig -> origin/gh/etaf/179/orig 2025-12-04T09:16:47.3282836Z * [new branch] gh/etaf/180/base -> origin/gh/etaf/180/base 2025-12-04T09:16:47.3284761Z * [new branch] gh/etaf/180/head -> origin/gh/etaf/180/head 2025-12-04T09:16:47.3286974Z * [new branch] gh/etaf/180/orig -> origin/gh/etaf/180/orig 2025-12-04T09:16:47.3291379Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-12-04T09:16:47.3293034Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-12-04T09:16:47.3296171Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-12-04T09:16:47.3297959Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-12-04T09:16:47.3300432Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-12-04T09:16:47.3302268Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-12-04T09:16:47.3304806Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-12-04T09:16:47.3306636Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-12-04T09:16:47.3309803Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-12-04T09:16:47.3311743Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-12-04T09:16:47.3313752Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-12-04T09:16:47.3316114Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-12-04T09:16:47.3317916Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-12-04T09:16:47.3319904Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-12-04T09:16:47.3322388Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-12-04T09:16:47.3324176Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-12-04T09:16:47.3326187Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-12-04T09:16:47.3328853Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-12-04T09:16:47.3330641Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-12-04T09:16:47.3332452Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-12-04T09:16:47.3334962Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-12-04T09:16:47.3337106Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-12-04T09:16:47.3338834Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-12-04T09:16:47.3341301Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-12-04T09:16:47.3343129Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-12-04T09:16:47.3345033Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-12-04T09:16:47.3347520Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-12-04T09:16:47.3349387Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-12-04T09:16:47.3351346Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-12-04T09:16:47.3353948Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-12-04T09:16:47.3355869Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-12-04T09:16:47.3357728Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-12-04T09:16:47.3360181Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-12-04T09:16:47.3361999Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-12-04T09:16:47.3363848Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-12-04T09:16:47.3366394Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-12-04T09:16:47.3368314Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-12-04T09:16:47.3370325Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-12-04T09:16:47.3372834Z * [new branch] gh/ezyang/3175/base -> origin/gh/ezyang/3175/base 2025-12-04T09:16:47.3374652Z * [new branch] gh/ezyang/3175/head -> origin/gh/ezyang/3175/head 2025-12-04T09:16:47.3376481Z * [new branch] gh/ezyang/3175/orig -> origin/gh/ezyang/3175/orig 2025-12-04T09:16:47.3378977Z * [new branch] gh/ezyang/3182/base -> origin/gh/ezyang/3182/base 2025-12-04T09:16:47.3380797Z * [new branch] gh/ezyang/3182/head -> origin/gh/ezyang/3182/head 2025-12-04T09:16:47.3382603Z * [new branch] gh/ezyang/3182/orig -> origin/gh/ezyang/3182/orig 2025-12-04T09:16:47.3385135Z * [new branch] gh/ezyang/3185/base -> origin/gh/ezyang/3185/base 2025-12-04T09:16:47.3387058Z * [new branch] gh/ezyang/3185/head -> origin/gh/ezyang/3185/head 2025-12-04T09:16:47.3388738Z * [new branch] gh/ezyang/3185/orig -> origin/gh/ezyang/3185/orig 2025-12-04T09:16:47.3391217Z * [new branch] gh/ezyang/3189/base -> origin/gh/ezyang/3189/base 2025-12-04T09:16:47.3393098Z * [new branch] gh/ezyang/3189/head -> origin/gh/ezyang/3189/head 2025-12-04T09:16:47.3394983Z * [new branch] gh/ezyang/3189/orig -> origin/gh/ezyang/3189/orig 2025-12-04T09:16:47.3397510Z * [new branch] gh/ezyang/3191/base -> origin/gh/ezyang/3191/base 2025-12-04T09:16:47.3399364Z * [new branch] gh/ezyang/3191/head -> origin/gh/ezyang/3191/head 2025-12-04T09:16:47.3401190Z * [new branch] gh/ezyang/3191/orig -> origin/gh/ezyang/3191/orig 2025-12-04T09:16:47.3404306Z * [new branch] gh/ezyang/3192/base -> origin/gh/ezyang/3192/base 2025-12-04T09:16:47.3406167Z * [new branch] gh/ezyang/3192/head -> origin/gh/ezyang/3192/head 2025-12-04T09:16:47.3408148Z * [new branch] gh/ezyang/3192/orig -> origin/gh/ezyang/3192/orig 2025-12-04T09:16:47.3410702Z * [new branch] gh/ezyang/3193/base -> origin/gh/ezyang/3193/base 2025-12-04T09:16:47.3412536Z * [new branch] gh/ezyang/3193/head -> origin/gh/ezyang/3193/head 2025-12-04T09:16:47.3414392Z * [new branch] gh/ezyang/3193/orig -> origin/gh/ezyang/3193/orig 2025-12-04T09:16:47.3417015Z * [new branch] gh/ezyang/3194/base -> origin/gh/ezyang/3194/base 2025-12-04T09:16:47.3418903Z * [new branch] gh/ezyang/3194/head -> origin/gh/ezyang/3194/head 2025-12-04T09:16:47.3420852Z * [new branch] gh/ezyang/3194/orig -> origin/gh/ezyang/3194/orig 2025-12-04T09:16:47.3423733Z * [new branch] gh/ezyang/3195/base -> origin/gh/ezyang/3195/base 2025-12-04T09:16:47.3425572Z * [new branch] gh/ezyang/3195/head -> origin/gh/ezyang/3195/head 2025-12-04T09:16:47.3427332Z * [new branch] gh/ezyang/3195/orig -> origin/gh/ezyang/3195/orig 2025-12-04T09:16:47.3429988Z * [new branch] gh/ezyang/3196/base -> origin/gh/ezyang/3196/base 2025-12-04T09:16:47.3431834Z * [new branch] gh/ezyang/3196/head -> origin/gh/ezyang/3196/head 2025-12-04T09:16:47.3433754Z * [new branch] gh/ezyang/3196/orig -> origin/gh/ezyang/3196/orig 2025-12-04T09:16:47.3437259Z * [new branch] gh/ezyang/3197/base -> origin/gh/ezyang/3197/base 2025-12-04T09:16:47.3441071Z * [new branch] gh/ezyang/3197/head -> origin/gh/ezyang/3197/head 2025-12-04T09:16:47.3442882Z * [new branch] gh/ezyang/3197/orig -> origin/gh/ezyang/3197/orig 2025-12-04T09:16:47.3445441Z * [new branch] gh/ezyang/3198/base -> origin/gh/ezyang/3198/base 2025-12-04T09:16:47.3447332Z * [new branch] gh/ezyang/3198/head -> origin/gh/ezyang/3198/head 2025-12-04T09:16:47.3449511Z * [new branch] gh/ezyang/3198/orig -> origin/gh/ezyang/3198/orig 2025-12-04T09:16:47.3452056Z * [new branch] gh/ezyang/3199/base -> origin/gh/ezyang/3199/base 2025-12-04T09:16:47.3453864Z * [new branch] gh/ezyang/3199/head -> origin/gh/ezyang/3199/head 2025-12-04T09:16:47.3455747Z * [new branch] gh/ezyang/3199/orig -> origin/gh/ezyang/3199/orig 2025-12-04T09:16:47.3458321Z * [new branch] gh/ezyang/3200/base -> origin/gh/ezyang/3200/base 2025-12-04T09:16:47.3460102Z * [new branch] gh/ezyang/3200/head -> origin/gh/ezyang/3200/head 2025-12-04T09:16:47.3462293Z * [new branch] gh/ezyang/3200/orig -> origin/gh/ezyang/3200/orig 2025-12-04T09:16:47.3465806Z * [new branch] gh/ezyang/3201/base -> origin/gh/ezyang/3201/base 2025-12-04T09:16:47.3468450Z * [new branch] gh/ezyang/3201/head -> origin/gh/ezyang/3201/head 2025-12-04T09:16:47.3470074Z * [new branch] gh/ezyang/3201/orig -> origin/gh/ezyang/3201/orig 2025-12-04T09:16:47.3472610Z * [new branch] gh/ezyang/3202/base -> origin/gh/ezyang/3202/base 2025-12-04T09:16:47.3474460Z * [new branch] gh/ezyang/3202/head -> origin/gh/ezyang/3202/head 2025-12-04T09:16:47.3476333Z * [new branch] gh/ezyang/3202/orig -> origin/gh/ezyang/3202/orig 2025-12-04T09:16:47.3478963Z * [new branch] gh/ezyang/3203/base -> origin/gh/ezyang/3203/base 2025-12-04T09:16:47.3480775Z * [new branch] gh/ezyang/3203/head -> origin/gh/ezyang/3203/head 2025-12-04T09:16:47.3482684Z * [new branch] gh/ezyang/3203/orig -> origin/gh/ezyang/3203/orig 2025-12-04T09:16:47.3485308Z * [new branch] gh/ezyang/3204/base -> origin/gh/ezyang/3204/base 2025-12-04T09:16:47.3487140Z * [new branch] gh/ezyang/3204/head -> origin/gh/ezyang/3204/head 2025-12-04T09:16:47.3489216Z * [new branch] gh/ezyang/3204/orig -> origin/gh/ezyang/3204/orig 2025-12-04T09:16:47.3491764Z * [new branch] gh/ezyang/3205/base -> origin/gh/ezyang/3205/base 2025-12-04T09:16:47.3493631Z * [new branch] gh/ezyang/3205/head -> origin/gh/ezyang/3205/head 2025-12-04T09:16:47.3495437Z * [new branch] gh/ezyang/3205/orig -> origin/gh/ezyang/3205/orig 2025-12-04T09:16:47.3497926Z * [new branch] gh/ezyang/3206/base -> origin/gh/ezyang/3206/base 2025-12-04T09:16:47.3499773Z * [new branch] gh/ezyang/3206/head -> origin/gh/ezyang/3206/head 2025-12-04T09:16:47.3501691Z * [new branch] gh/ezyang/3206/orig -> origin/gh/ezyang/3206/orig 2025-12-04T09:16:47.3504750Z * [new branch] gh/ezyang/3207/base -> origin/gh/ezyang/3207/base 2025-12-04T09:16:47.3506585Z * [new branch] gh/ezyang/3207/head -> origin/gh/ezyang/3207/head 2025-12-04T09:16:47.3508402Z * [new branch] gh/ezyang/3207/orig -> origin/gh/ezyang/3207/orig 2025-12-04T09:16:47.3510996Z * [new branch] gh/ezyang/3208/base -> origin/gh/ezyang/3208/base 2025-12-04T09:16:47.3512848Z * [new branch] gh/ezyang/3208/head -> origin/gh/ezyang/3208/head 2025-12-04T09:16:47.3514683Z * [new branch] gh/ezyang/3208/orig -> origin/gh/ezyang/3208/orig 2025-12-04T09:16:47.3517216Z * [new branch] gh/ezyang/3209/base -> origin/gh/ezyang/3209/base 2025-12-04T09:16:47.3519080Z * [new branch] gh/ezyang/3209/head -> origin/gh/ezyang/3209/head 2025-12-04T09:16:47.3520917Z * [new branch] gh/ezyang/3209/orig -> origin/gh/ezyang/3209/orig 2025-12-04T09:16:47.3523936Z * [new branch] gh/fadara01/3/base -> origin/gh/fadara01/3/base 2025-12-04T09:16:47.3525865Z * [new branch] gh/fadara01/3/head -> origin/gh/fadara01/3/head 2025-12-04T09:16:47.3527785Z * [new branch] gh/fadara01/3/orig -> origin/gh/fadara01/3/orig 2025-12-04T09:16:47.3530511Z * [new branch] gh/fadara01/5/base -> origin/gh/fadara01/5/base 2025-12-04T09:16:47.3532285Z * [new branch] gh/fadara01/5/head -> origin/gh/fadara01/5/head 2025-12-04T09:16:47.3534085Z * [new branch] gh/fadara01/5/orig -> origin/gh/fadara01/5/orig 2025-12-04T09:16:47.3536875Z * [new branch] gh/fadara01/6/base -> origin/gh/fadara01/6/base 2025-12-04T09:16:47.3538683Z * [new branch] gh/fadara01/6/head -> origin/gh/fadara01/6/head 2025-12-04T09:16:47.3540535Z * [new branch] gh/fadara01/6/orig -> origin/gh/fadara01/6/orig 2025-12-04T09:16:47.3543872Z * [new branch] gh/fadara01/7/base -> origin/gh/fadara01/7/base 2025-12-04T09:16:47.3545403Z * [new branch] gh/fadara01/7/head -> origin/gh/fadara01/7/head 2025-12-04T09:16:47.3547484Z * [new branch] gh/fadara01/7/orig -> origin/gh/fadara01/7/orig 2025-12-04T09:16:47.3550338Z * [new branch] gh/fadara01/8/base -> origin/gh/fadara01/8/base 2025-12-04T09:16:47.3551742Z * [new branch] gh/fadara01/8/head -> origin/gh/fadara01/8/head 2025-12-04T09:16:47.3555483Z * [new branch] gh/fadara01/8/orig -> origin/gh/fadara01/8/orig 2025-12-04T09:16:47.3556587Z * [new branch] gh/fadara01/9/base -> origin/gh/fadara01/9/base 2025-12-04T09:16:47.3558359Z * [new branch] gh/fadara01/9/head -> origin/gh/fadara01/9/head 2025-12-04T09:16:47.3560497Z * [new branch] gh/fadara01/9/orig -> origin/gh/fadara01/9/orig 2025-12-04T09:16:47.3563596Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-12-04T09:16:47.3565300Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-12-04T09:16:47.3566940Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-12-04T09:16:47.3570115Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-12-04T09:16:47.3571603Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-12-04T09:16:47.3573508Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-12-04T09:16:47.3576586Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-12-04T09:16:47.3577930Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-12-04T09:16:47.3579696Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-12-04T09:16:47.3582662Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-12-04T09:16:47.3584092Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-12-04T09:16:47.3585879Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-12-04T09:16:47.3588822Z * [new branch] gh/fduwjj/226/base -> origin/gh/fduwjj/226/base 2025-12-04T09:16:47.3590304Z * [new branch] gh/fduwjj/226/head -> origin/gh/fduwjj/226/head 2025-12-04T09:16:47.3592408Z * [new branch] gh/fduwjj/226/orig -> origin/gh/fduwjj/226/orig 2025-12-04T09:16:47.3595141Z * [new branch] gh/fduwjj/229/base -> origin/gh/fduwjj/229/base 2025-12-04T09:16:47.3596656Z * [new branch] gh/fduwjj/229/head -> origin/gh/fduwjj/229/head 2025-12-04T09:16:47.3598517Z * [new branch] gh/fduwjj/229/orig -> origin/gh/fduwjj/229/orig 2025-12-04T09:16:47.3601826Z * [new branch] gh/fduwjj/233/base -> origin/gh/fduwjj/233/base 2025-12-04T09:16:47.3603154Z * [new branch] gh/fduwjj/233/head -> origin/gh/fduwjj/233/head 2025-12-04T09:16:47.3605131Z * [new branch] gh/fduwjj/233/orig -> origin/gh/fduwjj/233/orig 2025-12-04T09:16:47.3608023Z * [new branch] gh/fduwjj/234/base -> origin/gh/fduwjj/234/base 2025-12-04T09:16:47.3609583Z * [new branch] gh/fduwjj/234/head -> origin/gh/fduwjj/234/head 2025-12-04T09:16:47.3611481Z * [new branch] gh/fduwjj/234/orig -> origin/gh/fduwjj/234/orig 2025-12-04T09:16:47.3614179Z * [new branch] gh/fduwjj/235/base -> origin/gh/fduwjj/235/base 2025-12-04T09:16:47.3615806Z * [new branch] gh/fduwjj/235/head -> origin/gh/fduwjj/235/head 2025-12-04T09:16:47.3617601Z * [new branch] gh/fduwjj/235/orig -> origin/gh/fduwjj/235/orig 2025-12-04T09:16:47.3620399Z * [new branch] gh/fduwjj/236/base -> origin/gh/fduwjj/236/base 2025-12-04T09:16:47.3621877Z * [new branch] gh/fduwjj/236/head -> origin/gh/fduwjj/236/head 2025-12-04T09:16:47.3623625Z * [new branch] gh/fduwjj/236/orig -> origin/gh/fduwjj/236/orig 2025-12-04T09:16:47.3626552Z * [new branch] gh/fduwjj/237/base -> origin/gh/fduwjj/237/base 2025-12-04T09:16:47.3628065Z * [new branch] gh/fduwjj/237/head -> origin/gh/fduwjj/237/head 2025-12-04T09:16:47.3629852Z * [new branch] gh/fduwjj/237/orig -> origin/gh/fduwjj/237/orig 2025-12-04T09:16:47.3632720Z * [new branch] gh/fduwjj/238/base -> origin/gh/fduwjj/238/base 2025-12-04T09:16:47.3634389Z * [new branch] gh/fduwjj/238/head -> origin/gh/fduwjj/238/head 2025-12-04T09:16:47.3636359Z * [new branch] gh/fduwjj/238/orig -> origin/gh/fduwjj/238/orig 2025-12-04T09:16:47.3641653Z * [new branch] gh/fduwjj/239/base -> origin/gh/fduwjj/239/base 2025-12-04T09:16:47.3643171Z * [new branch] gh/fduwjj/239/head -> origin/gh/fduwjj/239/head 2025-12-04T09:16:47.3644981Z * [new branch] gh/fduwjj/239/orig -> origin/gh/fduwjj/239/orig 2025-12-04T09:16:47.3648579Z * [new branch] gh/fegin/332/base -> origin/gh/fegin/332/base 2025-12-04T09:16:47.3650046Z * [new branch] gh/fegin/332/head -> origin/gh/fegin/332/head 2025-12-04T09:16:47.3652014Z * [new branch] gh/fegin/332/orig -> origin/gh/fegin/332/orig 2025-12-04T09:16:47.3654732Z * [new branch] gh/fegin/333/base -> origin/gh/fegin/333/base 2025-12-04T09:16:47.3656270Z * [new branch] gh/fegin/333/head -> origin/gh/fegin/333/head 2025-12-04T09:16:47.3658488Z * [new branch] gh/fegin/333/orig -> origin/gh/fegin/333/orig 2025-12-04T09:16:47.3661042Z * [new branch] gh/fegin/334/base -> origin/gh/fegin/334/base 2025-12-04T09:16:47.3662628Z * [new branch] gh/fegin/334/head -> origin/gh/fegin/334/head 2025-12-04T09:16:47.3664857Z * [new branch] gh/fegin/334/orig -> origin/gh/fegin/334/orig 2025-12-04T09:16:47.3667409Z * [new branch] gh/fegin/335/base -> origin/gh/fegin/335/base 2025-12-04T09:16:47.3668990Z * [new branch] gh/fegin/335/head -> origin/gh/fegin/335/head 2025-12-04T09:16:47.3670966Z * [new branch] gh/fegin/335/orig -> origin/gh/fegin/335/orig 2025-12-04T09:16:47.3674357Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-12-04T09:16:47.3675755Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-12-04T09:16:47.3678657Z * [new branch] gh/fffrog/177/base -> origin/gh/fffrog/177/base 2025-12-04T09:16:47.3680086Z * [new branch] gh/fffrog/177/head -> origin/gh/fffrog/177/head 2025-12-04T09:16:47.3682351Z * [new branch] gh/fffrog/177/orig -> origin/gh/fffrog/177/orig 2025-12-04T09:16:47.3684867Z * [new branch] gh/fffrog/178/base -> origin/gh/fffrog/178/base 2025-12-04T09:16:47.3686385Z * [new branch] gh/fffrog/178/head -> origin/gh/fffrog/178/head 2025-12-04T09:16:47.3689419Z * [new branch] gh/fffrog/178/orig -> origin/gh/fffrog/178/orig 2025-12-04T09:16:47.3691893Z * [new branch] gh/fffrog/181/base -> origin/gh/fffrog/181/base 2025-12-04T09:16:47.3693275Z * [new branch] gh/fffrog/181/head -> origin/gh/fffrog/181/head 2025-12-04T09:16:47.3695330Z * [new branch] gh/fffrog/181/orig -> origin/gh/fffrog/181/orig 2025-12-04T09:16:47.3697967Z * [new branch] gh/fffrog/183/base -> origin/gh/fffrog/183/base 2025-12-04T09:16:47.3699584Z * [new branch] gh/fffrog/183/head -> origin/gh/fffrog/183/head 2025-12-04T09:16:47.3701390Z * [new branch] gh/fffrog/183/orig -> origin/gh/fffrog/183/orig 2025-12-04T09:16:47.3704633Z * [new branch] gh/fxdawnn/10/base -> origin/gh/fxdawnn/10/base 2025-12-04T09:16:47.3707009Z * [new branch] gh/fxdawnn/10/head -> origin/gh/fxdawnn/10/head 2025-12-04T09:16:47.3708158Z * [new branch] gh/fxdawnn/10/orig -> origin/gh/fxdawnn/10/orig 2025-12-04T09:16:47.3711320Z * [new branch] gh/fxdawnn/11/base -> origin/gh/fxdawnn/11/base 2025-12-04T09:16:47.3712766Z * [new branch] gh/fxdawnn/11/head -> origin/gh/fxdawnn/11/head 2025-12-04T09:16:47.3715258Z * [new branch] gh/fxdawnn/11/orig -> origin/gh/fxdawnn/11/orig 2025-12-04T09:16:47.3717575Z * [new branch] gh/fxdawnn/12/base -> origin/gh/fxdawnn/12/base 2025-12-04T09:16:47.3719173Z * [new branch] gh/fxdawnn/12/head -> origin/gh/fxdawnn/12/head 2025-12-04T09:16:47.3721239Z * [new branch] gh/fxdawnn/12/orig -> origin/gh/fxdawnn/12/orig 2025-12-04T09:16:47.3723824Z * [new branch] gh/fxdawnn/13/base -> origin/gh/fxdawnn/13/base 2025-12-04T09:16:47.3725362Z * [new branch] gh/fxdawnn/13/head -> origin/gh/fxdawnn/13/head 2025-12-04T09:16:47.3727749Z * [new branch] gh/fxdawnn/13/orig -> origin/gh/fxdawnn/13/orig 2025-12-04T09:16:47.3730601Z * [new branch] gh/fxdawnn/14/base -> origin/gh/fxdawnn/14/base 2025-12-04T09:16:47.3732004Z * [new branch] gh/fxdawnn/14/head -> origin/gh/fxdawnn/14/head 2025-12-04T09:16:47.3733941Z * [new branch] gh/fxdawnn/14/orig -> origin/gh/fxdawnn/14/orig 2025-12-04T09:16:47.3736665Z * [new branch] gh/fxdawnn/15/base -> origin/gh/fxdawnn/15/base 2025-12-04T09:16:47.3738673Z * [new branch] gh/fxdawnn/15/head -> origin/gh/fxdawnn/15/head 2025-12-04T09:16:47.3740257Z * [new branch] gh/fxdawnn/15/orig -> origin/gh/fxdawnn/15/orig 2025-12-04T09:16:47.3743206Z * [new branch] gh/fxdawnn/6/base -> origin/gh/fxdawnn/6/base 2025-12-04T09:16:47.3744635Z * [new branch] gh/fxdawnn/6/head -> origin/gh/fxdawnn/6/head 2025-12-04T09:16:47.3746911Z * [new branch] gh/fxdawnn/6/orig -> origin/gh/fxdawnn/6/orig 2025-12-04T09:16:47.3749348Z * [new branch] gh/fxdawnn/7/base -> origin/gh/fxdawnn/7/base 2025-12-04T09:16:47.3751050Z * [new branch] gh/fxdawnn/7/head -> origin/gh/fxdawnn/7/head 2025-12-04T09:16:47.3753004Z * [new branch] gh/fxdawnn/7/orig -> origin/gh/fxdawnn/7/orig 2025-12-04T09:16:47.3755760Z * [new branch] gh/fxdawnn/9/base -> origin/gh/fxdawnn/9/base 2025-12-04T09:16:47.3758031Z * [new branch] gh/fxdawnn/9/head -> origin/gh/fxdawnn/9/head 2025-12-04T09:16:47.3759663Z * [new branch] gh/fxdawnn/9/orig -> origin/gh/fxdawnn/9/orig 2025-12-04T09:16:47.3763099Z * [new branch] gh/galv/1/base -> origin/gh/galv/1/base 2025-12-04T09:16:47.3764602Z * [new branch] gh/galv/1/head -> origin/gh/galv/1/head 2025-12-04T09:16:47.3766722Z * [new branch] gh/galv/1/orig -> origin/gh/galv/1/orig 2025-12-04T09:16:47.3769937Z * [new branch] gh/galv/2/base -> origin/gh/galv/2/base 2025-12-04T09:16:47.3771766Z * [new branch] gh/galv/2/head -> origin/gh/galv/2/head 2025-12-04T09:16:47.3773643Z * [new branch] gh/galv/2/orig -> origin/gh/galv/2/orig 2025-12-04T09:16:47.3776402Z * [new branch] gh/galv/3/base -> origin/gh/galv/3/base 2025-12-04T09:16:47.3778100Z * [new branch] gh/galv/3/head -> origin/gh/galv/3/head 2025-12-04T09:16:47.3780080Z * [new branch] gh/galv/3/orig -> origin/gh/galv/3/orig 2025-12-04T09:16:47.3783206Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-12-04T09:16:47.3785072Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-12-04T09:16:47.3786934Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-12-04T09:16:47.3789444Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-12-04T09:16:47.3791282Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-12-04T09:16:47.3793076Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-12-04T09:16:47.3795612Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-12-04T09:16:47.3798031Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-12-04T09:16:47.3799878Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-12-04T09:16:47.3802504Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-12-04T09:16:47.3804363Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-12-04T09:16:47.3806213Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-12-04T09:16:47.3808966Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-12-04T09:16:47.3810849Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-12-04T09:16:47.3812669Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-12-04T09:16:47.3815186Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-12-04T09:16:47.3817068Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-12-04T09:16:47.3818926Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-12-04T09:16:47.3821393Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-12-04T09:16:47.3823312Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-12-04T09:16:47.3825069Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-12-04T09:16:47.3827729Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-12-04T09:16:47.3829587Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-12-04T09:16:47.3831427Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-12-04T09:16:47.3833893Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-12-04T09:16:47.3835918Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-12-04T09:16:47.3837933Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-12-04T09:16:47.3840409Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-12-04T09:16:47.3842248Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-12-04T09:16:47.3844033Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-12-04T09:16:47.3846623Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-12-04T09:16:47.3848622Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-12-04T09:16:47.3850564Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-12-04T09:16:47.3853049Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-12-04T09:16:47.3854857Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-12-04T09:16:47.3856740Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-12-04T09:16:47.3859261Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-12-04T09:16:47.3861677Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-12-04T09:16:47.3863611Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-12-04T09:16:47.3866130Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-12-04T09:16:47.3867929Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-12-04T09:16:47.3869737Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-12-04T09:16:47.3872295Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-12-04T09:16:47.3874065Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-12-04T09:16:47.3875890Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-12-04T09:16:47.3878660Z * [new branch] gh/guangyey/228/base -> origin/gh/guangyey/228/base 2025-12-04T09:16:47.3880583Z * [new branch] gh/guangyey/228/head -> origin/gh/guangyey/228/head 2025-12-04T09:16:47.3882437Z * [new branch] gh/guangyey/228/orig -> origin/gh/guangyey/228/orig 2025-12-04T09:16:47.3885467Z * [new branch] gh/guangyey/230/base -> origin/gh/guangyey/230/base 2025-12-04T09:16:47.3887343Z * [new branch] gh/guangyey/230/head -> origin/gh/guangyey/230/head 2025-12-04T09:16:47.3889292Z * [new branch] gh/guangyey/230/orig -> origin/gh/guangyey/230/orig 2025-12-04T09:16:47.3891882Z * [new branch] gh/guangyey/231/base -> origin/gh/guangyey/231/base 2025-12-04T09:16:47.3893757Z * [new branch] gh/guangyey/231/head -> origin/gh/guangyey/231/head 2025-12-04T09:16:47.3895598Z * [new branch] gh/guangyey/231/orig -> origin/gh/guangyey/231/orig 2025-12-04T09:16:47.3898169Z * [new branch] gh/guangyey/232/base -> origin/gh/guangyey/232/base 2025-12-04T09:16:47.3900003Z * [new branch] gh/guangyey/232/head -> origin/gh/guangyey/232/head 2025-12-04T09:16:47.3901821Z * [new branch] gh/guangyey/232/orig -> origin/gh/guangyey/232/orig 2025-12-04T09:16:47.3904464Z * [new branch] gh/guangyey/233/base -> origin/gh/guangyey/233/base 2025-12-04T09:16:47.3906260Z * [new branch] gh/guangyey/233/head -> origin/gh/guangyey/233/head 2025-12-04T09:16:47.3908121Z * [new branch] gh/guangyey/233/orig -> origin/gh/guangyey/233/orig 2025-12-04T09:16:47.3910679Z * [new branch] gh/guangyey/234/base -> origin/gh/guangyey/234/base 2025-12-04T09:16:47.3912584Z * [new branch] gh/guangyey/234/head -> origin/gh/guangyey/234/head 2025-12-04T09:16:47.3914422Z * [new branch] gh/guangyey/234/orig -> origin/gh/guangyey/234/orig 2025-12-04T09:16:47.3917016Z * [new branch] gh/guangyey/235/base -> origin/gh/guangyey/235/base 2025-12-04T09:16:47.3918808Z * [new branch] gh/guangyey/235/head -> origin/gh/guangyey/235/head 2025-12-04T09:16:47.3920614Z * [new branch] gh/guangyey/235/orig -> origin/gh/guangyey/235/orig 2025-12-04T09:16:47.3923177Z * [new branch] gh/guangyey/236/base -> origin/gh/guangyey/236/base 2025-12-04T09:16:47.3925152Z * [new branch] gh/guangyey/236/head -> origin/gh/guangyey/236/head 2025-12-04T09:16:47.3926931Z * [new branch] gh/guangyey/236/orig -> origin/gh/guangyey/236/orig 2025-12-04T09:16:47.3929797Z * [new branch] gh/guangyey/237/base -> origin/gh/guangyey/237/base 2025-12-04T09:16:47.3931621Z * [new branch] gh/guangyey/237/head -> origin/gh/guangyey/237/head 2025-12-04T09:16:47.3933463Z * [new branch] gh/guangyey/237/orig -> origin/gh/guangyey/237/orig 2025-12-04T09:16:47.3936115Z * [new branch] gh/guangyey/238/base -> origin/gh/guangyey/238/base 2025-12-04T09:16:47.3938028Z * [new branch] gh/guangyey/238/head -> origin/gh/guangyey/238/head 2025-12-04T09:16:47.3940579Z * [new branch] gh/guangyey/239/base -> origin/gh/guangyey/239/base 2025-12-04T09:16:47.3942448Z * [new branch] gh/guangyey/239/head -> origin/gh/guangyey/239/head 2025-12-04T09:16:47.3944262Z * [new branch] gh/guangyey/239/orig -> origin/gh/guangyey/239/orig 2025-12-04T09:16:47.3946861Z * [new branch] gh/guangyey/240/base -> origin/gh/guangyey/240/base 2025-12-04T09:16:47.3948709Z * [new branch] gh/guangyey/240/head -> origin/gh/guangyey/240/head 2025-12-04T09:16:47.3950560Z * [new branch] gh/guangyey/240/orig -> origin/gh/guangyey/240/orig 2025-12-04T09:16:47.3953190Z * [new branch] gh/guangyey/241/base -> origin/gh/guangyey/241/base 2025-12-04T09:16:47.3955087Z * [new branch] gh/guangyey/241/head -> origin/gh/guangyey/241/head 2025-12-04T09:16:47.3966524Z * [new branch] gh/guangyey/241/orig -> origin/gh/guangyey/241/orig 2025-12-04T09:16:47.3967222Z * [new branch] gh/guangyey/242/base -> origin/gh/guangyey/242/base 2025-12-04T09:16:47.3968034Z * [new branch] gh/guangyey/242/head -> origin/gh/guangyey/242/head 2025-12-04T09:16:47.3968777Z * [new branch] gh/guangyey/242/orig -> origin/gh/guangyey/242/orig 2025-12-04T09:16:47.3969492Z * [new branch] gh/guangyey/243/base -> origin/gh/guangyey/243/base 2025-12-04T09:16:47.3970205Z * [new branch] gh/guangyey/243/head -> origin/gh/guangyey/243/head 2025-12-04T09:16:47.3970769Z * [new branch] gh/guangyey/243/orig -> origin/gh/guangyey/243/orig 2025-12-04T09:16:47.3972163Z * [new branch] gh/guangyey/244/base -> origin/gh/guangyey/244/base 2025-12-04T09:16:47.3974206Z * [new branch] gh/guangyey/244/head -> origin/gh/guangyey/244/head 2025-12-04T09:16:47.3975984Z * [new branch] gh/guangyey/244/orig -> origin/gh/guangyey/244/orig 2025-12-04T09:16:47.3978491Z * [new branch] gh/guangyey/245/base -> origin/gh/guangyey/245/base 2025-12-04T09:16:47.3980481Z * [new branch] gh/guangyey/245/head -> origin/gh/guangyey/245/head 2025-12-04T09:16:47.3982348Z * [new branch] gh/guangyey/245/orig -> origin/gh/guangyey/245/orig 2025-12-04T09:16:47.3984876Z * [new branch] gh/guangyey/246/base -> origin/gh/guangyey/246/base 2025-12-04T09:16:47.3986696Z * [new branch] gh/guangyey/246/head -> origin/gh/guangyey/246/head 2025-12-04T09:16:47.3988511Z * [new branch] gh/guangyey/246/orig -> origin/gh/guangyey/246/orig 2025-12-04T09:16:47.3991104Z * [new branch] gh/guangyey/247/base -> origin/gh/guangyey/247/base 2025-12-04T09:16:47.3992931Z * [new branch] gh/guangyey/247/head -> origin/gh/guangyey/247/head 2025-12-04T09:16:47.3994741Z * [new branch] gh/guangyey/247/orig -> origin/gh/guangyey/247/orig 2025-12-04T09:16:47.3997463Z * [new branch] gh/guangyey/248/base -> origin/gh/guangyey/248/base 2025-12-04T09:16:47.3999447Z * [new branch] gh/guangyey/248/head -> origin/gh/guangyey/248/head 2025-12-04T09:16:47.4001118Z * [new branch] gh/guangyey/248/orig -> origin/gh/guangyey/248/orig 2025-12-04T09:16:47.4003672Z * [new branch] gh/guangyey/249/base -> origin/gh/guangyey/249/base 2025-12-04T09:16:47.4005726Z * [new branch] gh/guangyey/249/head -> origin/gh/guangyey/249/head 2025-12-04T09:16:47.4007655Z * [new branch] gh/guangyey/249/orig -> origin/gh/guangyey/249/orig 2025-12-04T09:16:47.4010252Z * [new branch] gh/guangyey/250/base -> origin/gh/guangyey/250/base 2025-12-04T09:16:47.4012141Z * [new branch] gh/guangyey/250/head -> origin/gh/guangyey/250/head 2025-12-04T09:16:47.4013922Z * [new branch] gh/guangyey/250/orig -> origin/gh/guangyey/250/orig 2025-12-04T09:16:47.4016419Z * [new branch] gh/guangyey/251/base -> origin/gh/guangyey/251/base 2025-12-04T09:16:47.4018265Z * [new branch] gh/guangyey/251/head -> origin/gh/guangyey/251/head 2025-12-04T09:16:47.4020172Z * [new branch] gh/guangyey/251/orig -> origin/gh/guangyey/251/orig 2025-12-04T09:16:47.4022748Z * [new branch] gh/guangyey/252/base -> origin/gh/guangyey/252/base 2025-12-04T09:16:47.4024651Z * [new branch] gh/guangyey/252/head -> origin/gh/guangyey/252/head 2025-12-04T09:16:47.4026494Z * [new branch] gh/guangyey/252/orig -> origin/gh/guangyey/252/orig 2025-12-04T09:16:47.4029032Z * [new branch] gh/guangyey/253/base -> origin/gh/guangyey/253/base 2025-12-04T09:16:47.4030945Z * [new branch] gh/guangyey/253/head -> origin/gh/guangyey/253/head 2025-12-04T09:16:47.4032754Z * [new branch] gh/guangyey/253/orig -> origin/gh/guangyey/253/orig 2025-12-04T09:16:47.4035617Z * [new branch] gh/guangyey/254/base -> origin/gh/guangyey/254/base 2025-12-04T09:16:47.4039998Z * [new branch] gh/guangyey/254/head -> origin/gh/guangyey/254/head 2025-12-04T09:16:47.4041813Z * [new branch] gh/guangyey/254/orig -> origin/gh/guangyey/254/orig 2025-12-04T09:16:47.4044345Z * [new branch] gh/guangyey/255/base -> origin/gh/guangyey/255/base 2025-12-04T09:16:47.4046152Z * [new branch] gh/guangyey/255/head -> origin/gh/guangyey/255/head 2025-12-04T09:16:47.4048065Z * [new branch] gh/guangyey/255/orig -> origin/gh/guangyey/255/orig 2025-12-04T09:16:47.4051477Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-12-04T09:16:47.4053610Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-12-04T09:16:47.4055557Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-12-04T09:16:47.4059307Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-12-04T09:16:47.4061639Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-12-04T09:16:47.4064159Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-12-04T09:16:47.4067576Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-12-04T09:16:47.4071724Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-12-04T09:16:47.4075085Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-12-04T09:16:47.4076772Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-12-04T09:16:47.4078797Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-12-04T09:16:47.4080459Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-12-04T09:16:47.4083198Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-12-04T09:16:47.4084859Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-12-04T09:16:47.4086663Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-12-04T09:16:47.4089505Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-12-04T09:16:47.4091713Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-12-04T09:16:47.4093644Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-12-04T09:16:47.4096082Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-12-04T09:16:47.4097914Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-12-04T09:16:47.4099855Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-12-04T09:16:47.4102329Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-12-04T09:16:47.4104130Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-12-04T09:16:47.4105965Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-12-04T09:16:47.4108528Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-12-04T09:16:47.4110366Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-12-04T09:16:47.4112325Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-12-04T09:16:47.4114968Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-12-04T09:16:47.4116759Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-12-04T09:16:47.4118541Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-12-04T09:16:47.4121140Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-12-04T09:16:47.4122989Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-12-04T09:16:47.4124840Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-12-04T09:16:47.4127370Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-12-04T09:16:47.4129453Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-12-04T09:16:47.4131187Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-12-04T09:16:47.4133689Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-12-04T09:16:47.4135780Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-12-04T09:16:47.4137652Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-12-04T09:16:47.4140258Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-12-04T09:16:47.4142057Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-12-04T09:16:47.4143958Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-12-04T09:16:47.4146469Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-12-04T09:16:47.4148301Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-12-04T09:16:47.4150125Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-12-04T09:16:47.4153427Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-12-04T09:16:47.4155159Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-12-04T09:16:47.4157012Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-12-04T09:16:47.4160069Z * [new branch] gh/guilhermeleobas/253/base -> origin/gh/guilhermeleobas/253/base 2025-12-04T09:16:47.4161904Z * [new branch] gh/guilhermeleobas/253/head -> origin/gh/guilhermeleobas/253/head 2025-12-04T09:16:47.4163790Z * [new branch] gh/guilhermeleobas/253/orig -> origin/gh/guilhermeleobas/253/orig 2025-12-04T09:16:47.4166617Z * [new branch] gh/guilhermeleobas/254/base -> origin/gh/guilhermeleobas/254/base 2025-12-04T09:16:47.4168575Z * [new branch] gh/guilhermeleobas/254/head -> origin/gh/guilhermeleobas/254/head 2025-12-04T09:16:47.4170452Z * [new branch] gh/guilhermeleobas/254/orig -> origin/gh/guilhermeleobas/254/orig 2025-12-04T09:16:47.4173135Z * [new branch] gh/guilhermeleobas/255/base -> origin/gh/guilhermeleobas/255/base 2025-12-04T09:16:47.4174915Z * [new branch] gh/guilhermeleobas/255/head -> origin/gh/guilhermeleobas/255/head 2025-12-04T09:16:47.4177280Z * [new branch] gh/guilhermeleobas/255/orig -> origin/gh/guilhermeleobas/255/orig 2025-12-04T09:16:47.4179852Z * [new branch] gh/guilhermeleobas/256/base -> origin/gh/guilhermeleobas/256/base 2025-12-04T09:16:47.4181662Z * [new branch] gh/guilhermeleobas/256/head -> origin/gh/guilhermeleobas/256/head 2025-12-04T09:16:47.4183242Z * [new branch] gh/guilhermeleobas/256/orig -> origin/gh/guilhermeleobas/256/orig 2025-12-04T09:16:47.4186037Z * [new branch] gh/guilhermeleobas/257/base -> origin/gh/guilhermeleobas/257/base 2025-12-04T09:16:47.4187699Z * [new branch] gh/guilhermeleobas/257/head -> origin/gh/guilhermeleobas/257/head 2025-12-04T09:16:47.4189564Z * [new branch] gh/guilhermeleobas/257/orig -> origin/gh/guilhermeleobas/257/orig 2025-12-04T09:16:47.4192604Z * [new branch] gh/guilhermeleobas/258/base -> origin/gh/guilhermeleobas/258/base 2025-12-04T09:16:47.4194139Z * [new branch] gh/guilhermeleobas/258/head -> origin/gh/guilhermeleobas/258/head 2025-12-04T09:16:47.4195933Z * [new branch] gh/guilhermeleobas/258/orig -> origin/gh/guilhermeleobas/258/orig 2025-12-04T09:16:47.4198787Z * [new branch] gh/guilhermeleobas/259/base -> origin/gh/guilhermeleobas/259/base 2025-12-04T09:16:47.4200440Z * [new branch] gh/guilhermeleobas/259/head -> origin/gh/guilhermeleobas/259/head 2025-12-04T09:16:47.4202241Z * [new branch] gh/guilhermeleobas/259/orig -> origin/gh/guilhermeleobas/259/orig 2025-12-04T09:16:47.4205673Z * [new branch] gh/guilhermeleobas/260/base -> origin/gh/guilhermeleobas/260/base 2025-12-04T09:16:47.4207301Z * [new branch] gh/guilhermeleobas/260/head -> origin/gh/guilhermeleobas/260/head 2025-12-04T09:16:47.4209584Z * [new branch] gh/guilhermeleobas/260/orig -> origin/gh/guilhermeleobas/260/orig 2025-12-04T09:16:47.4212160Z * [new branch] gh/guilhermeleobas/261/base -> origin/gh/guilhermeleobas/261/base 2025-12-04T09:16:47.4213779Z * [new branch] gh/guilhermeleobas/261/head -> origin/gh/guilhermeleobas/261/head 2025-12-04T09:16:47.4216346Z * [new branch] gh/guilhermeleobas/261/orig -> origin/gh/guilhermeleobas/261/orig 2025-12-04T09:16:47.4219054Z * [new branch] gh/guilhermeleobas/262/base -> origin/gh/guilhermeleobas/262/base 2025-12-04T09:16:47.4220759Z * [new branch] gh/guilhermeleobas/262/head -> origin/gh/guilhermeleobas/262/head 2025-12-04T09:16:47.4222529Z * [new branch] gh/guilhermeleobas/262/orig -> origin/gh/guilhermeleobas/262/orig 2025-12-04T09:16:47.4226039Z * [new branch] gh/guilhermeleobas/263/base -> origin/gh/guilhermeleobas/263/base 2025-12-04T09:16:47.4227474Z * [new branch] gh/guilhermeleobas/263/head -> origin/gh/guilhermeleobas/263/head 2025-12-04T09:16:47.4229258Z * [new branch] gh/guilhermeleobas/263/orig -> origin/gh/guilhermeleobas/263/orig 2025-12-04T09:16:47.4232133Z * [new branch] gh/guilhermeleobas/264/base -> origin/gh/guilhermeleobas/264/base 2025-12-04T09:16:47.4233866Z * [new branch] gh/guilhermeleobas/264/head -> origin/gh/guilhermeleobas/264/head 2025-12-04T09:16:47.4236211Z * [new branch] gh/guilhermeleobas/264/orig -> origin/gh/guilhermeleobas/264/orig 2025-12-04T09:16:47.4238630Z * [new branch] gh/guilhermeleobas/265/base -> origin/gh/guilhermeleobas/265/base 2025-12-04T09:16:47.4240375Z * [new branch] gh/guilhermeleobas/265/head -> origin/gh/guilhermeleobas/265/head 2025-12-04T09:16:47.4242223Z * [new branch] gh/guilhermeleobas/265/orig -> origin/gh/guilhermeleobas/265/orig 2025-12-04T09:16:47.4245061Z * [new branch] gh/guilhermeleobas/266/base -> origin/gh/guilhermeleobas/266/base 2025-12-04T09:16:47.4246873Z * [new branch] gh/guilhermeleobas/266/head -> origin/gh/guilhermeleobas/266/head 2025-12-04T09:16:47.4248896Z * [new branch] gh/guilhermeleobas/266/orig -> origin/gh/guilhermeleobas/266/orig 2025-12-04T09:16:47.4251584Z * [new branch] gh/guilhermeleobas/267/base -> origin/gh/guilhermeleobas/267/base 2025-12-04T09:16:47.4253434Z * [new branch] gh/guilhermeleobas/267/head -> origin/gh/guilhermeleobas/267/head 2025-12-04T09:16:47.4255270Z * [new branch] gh/guilhermeleobas/267/orig -> origin/gh/guilhermeleobas/267/orig 2025-12-04T09:16:47.4258474Z * [new branch] gh/hameerabbasi/1/base -> origin/gh/hameerabbasi/1/base 2025-12-04T09:16:47.4260346Z * [new branch] gh/hameerabbasi/1/head -> origin/gh/hameerabbasi/1/head 2025-12-04T09:16:47.4262796Z * [new branch] gh/hameerabbasi/2/base -> origin/gh/hameerabbasi/2/base 2025-12-04T09:16:47.4264649Z * [new branch] gh/hameerabbasi/2/head -> origin/gh/hameerabbasi/2/head 2025-12-04T09:16:47.4266522Z * [new branch] gh/hameerabbasi/2/orig -> origin/gh/hameerabbasi/2/orig 2025-12-04T09:16:47.4269158Z * [new branch] gh/hameerabbasi/3/base -> origin/gh/hameerabbasi/3/base 2025-12-04T09:16:47.4270958Z * [new branch] gh/hameerabbasi/3/head -> origin/gh/hameerabbasi/3/head 2025-12-04T09:16:47.4272929Z * [new branch] gh/hameerabbasi/3/orig -> origin/gh/hameerabbasi/3/orig 2025-12-04T09:16:47.4275344Z * [new branch] gh/hameerabbasi/4/base -> origin/gh/hameerabbasi/4/base 2025-12-04T09:16:47.4277166Z * [new branch] gh/hameerabbasi/4/head -> origin/gh/hameerabbasi/4/head 2025-12-04T09:16:47.4278878Z * [new branch] gh/hameerabbasi/4/orig -> origin/gh/hameerabbasi/4/orig 2025-12-04T09:16:47.4281934Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-12-04T09:16:47.4284355Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-12-04T09:16:47.4286900Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-12-04T09:16:47.4289521Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-12-04T09:16:47.4292110Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-12-04T09:16:47.4294593Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-12-04T09:16:47.4297640Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-12-04T09:16:47.4299495Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-12-04T09:16:47.4302722Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-12-04T09:16:47.4304562Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-12-04T09:16:47.4307003Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-12-04T09:16:47.4308844Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-12-04T09:16:47.4310730Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-12-04T09:16:47.4313377Z * [new branch] gh/isuruf/158/base -> origin/gh/isuruf/158/base 2025-12-04T09:16:47.4315204Z * [new branch] gh/isuruf/158/head -> origin/gh/isuruf/158/head 2025-12-04T09:16:47.4317591Z * [new branch] gh/isuruf/159/base -> origin/gh/isuruf/159/base 2025-12-04T09:16:47.4319431Z * [new branch] gh/isuruf/159/head -> origin/gh/isuruf/159/head 2025-12-04T09:16:47.4321961Z * [new branch] gh/isuruf/160/base -> origin/gh/isuruf/160/base 2025-12-04T09:16:47.4323766Z * [new branch] gh/isuruf/160/head -> origin/gh/isuruf/160/head 2025-12-04T09:16:47.4325637Z * [new branch] gh/isuruf/160/orig -> origin/gh/isuruf/160/orig 2025-12-04T09:16:47.4328375Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-12-04T09:16:47.4330120Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-12-04T09:16:47.4331907Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-12-04T09:16:47.4335009Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-12-04T09:16:47.4337220Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-12-04T09:16:47.4339006Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-12-04T09:16:47.4341471Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-12-04T09:16:47.4343246Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-12-04T09:16:47.4345058Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-12-04T09:16:47.4348067Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-12-04T09:16:47.4349938Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-12-04T09:16:47.4351844Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-12-04T09:16:47.4355466Z * [new branch] gh/jamesjwu/198/base -> origin/gh/jamesjwu/198/base 2025-12-04T09:16:47.4357294Z * [new branch] gh/jamesjwu/198/head -> origin/gh/jamesjwu/198/head 2025-12-04T09:16:47.4359143Z * [new branch] gh/jamesjwu/198/orig -> origin/gh/jamesjwu/198/orig 2025-12-04T09:16:47.4361919Z * [new branch] gh/jamesjwu/207/base -> origin/gh/jamesjwu/207/base 2025-12-04T09:16:47.4363885Z * [new branch] gh/jamesjwu/207/head -> origin/gh/jamesjwu/207/head 2025-12-04T09:16:47.4365752Z * [new branch] gh/jamesjwu/207/orig -> origin/gh/jamesjwu/207/orig 2025-12-04T09:16:47.4368859Z * [new branch] gh/jamesjwu/208/base -> origin/gh/jamesjwu/208/base 2025-12-04T09:16:47.4370475Z * [new branch] gh/jamesjwu/208/head -> origin/gh/jamesjwu/208/head 2025-12-04T09:16:47.4372307Z * [new branch] gh/jamesjwu/208/orig -> origin/gh/jamesjwu/208/orig 2025-12-04T09:16:47.4374915Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-12-04T09:16:47.4376822Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-12-04T09:16:47.4379402Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-12-04T09:16:47.4381053Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-12-04T09:16:47.4383418Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-12-04T09:16:47.4385260Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-12-04T09:16:47.4387737Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-12-04T09:16:47.4389610Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-12-04T09:16:47.4392005Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-12-04T09:16:47.4393825Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-12-04T09:16:47.4396182Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-12-04T09:16:47.4397971Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-12-04T09:16:47.4400411Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-12-04T09:16:47.4402276Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-12-04T09:16:47.4404646Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-12-04T09:16:47.4406555Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-12-04T09:16:47.4409347Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-12-04T09:16:47.4411333Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-12-04T09:16:47.4413578Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-12-04T09:16:47.4415324Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-12-04T09:16:47.4418549Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-12-04T09:16:47.4419812Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-12-04T09:16:47.4422409Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-12-04T09:16:47.4424278Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-12-04T09:16:47.4427772Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-12-04T09:16:47.4429576Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-12-04T09:16:47.4432032Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-12-04T09:16:47.4433928Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-12-04T09:16:47.4438625Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-12-04T09:16:47.4440537Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-12-04T09:16:47.4442399Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-12-04T09:16:47.4444796Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-12-04T09:16:47.4446637Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-12-04T09:16:47.4448571Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-12-04T09:16:47.4451379Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-12-04T09:16:47.4453337Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-12-04T09:16:47.4455073Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-12-04T09:16:47.4457582Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-12-04T09:16:47.4459696Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-12-04T09:16:47.4461433Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-12-04T09:16:47.4464208Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-12-04T09:16:47.4466080Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-12-04T09:16:47.4468403Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-12-04T09:16:47.4470151Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-12-04T09:16:47.4472655Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-12-04T09:16:47.4474515Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-12-04T09:16:47.4476871Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-12-04T09:16:47.4478638Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-12-04T09:16:47.4481140Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-12-04T09:16:47.4483115Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-12-04T09:16:47.4485026Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-12-04T09:16:47.4487543Z * [new branch] gh/janeyx99/315/base -> origin/gh/janeyx99/315/base 2025-12-04T09:16:47.4489496Z * [new branch] gh/janeyx99/315/head -> origin/gh/janeyx99/315/head 2025-12-04T09:16:47.4491422Z * [new branch] gh/janeyx99/315/orig -> origin/gh/janeyx99/315/orig 2025-12-04T09:16:47.4493878Z * [new branch] gh/janeyx99/316/base -> origin/gh/janeyx99/316/base 2025-12-04T09:16:47.4495665Z * [new branch] gh/janeyx99/316/head -> origin/gh/janeyx99/316/head 2025-12-04T09:16:47.4497505Z * [new branch] gh/janeyx99/316/orig -> origin/gh/janeyx99/316/orig 2025-12-04T09:16:47.4500751Z * [new branch] gh/janeyx99/317/base -> origin/gh/janeyx99/317/base 2025-12-04T09:16:47.4502537Z * [new branch] gh/janeyx99/317/head -> origin/gh/janeyx99/317/head 2025-12-04T09:16:47.4504292Z * [new branch] gh/janeyx99/317/orig -> origin/gh/janeyx99/317/orig 2025-12-04T09:16:47.4506917Z * [new branch] gh/janeyx99/325/base -> origin/gh/janeyx99/325/base 2025-12-04T09:16:47.4509036Z * [new branch] gh/janeyx99/325/head -> origin/gh/janeyx99/325/head 2025-12-04T09:16:47.4510830Z * [new branch] gh/janeyx99/325/orig -> origin/gh/janeyx99/325/orig 2025-12-04T09:16:47.4513321Z * [new branch] gh/janeyx99/327/base -> origin/gh/janeyx99/327/base 2025-12-04T09:16:47.4515197Z * [new branch] gh/janeyx99/327/head -> origin/gh/janeyx99/327/head 2025-12-04T09:16:47.4517047Z * [new branch] gh/janeyx99/327/orig -> origin/gh/janeyx99/327/orig 2025-12-04T09:16:47.4519487Z * [new branch] gh/janeyx99/328/base -> origin/gh/janeyx99/328/base 2025-12-04T09:16:47.4521379Z * [new branch] gh/janeyx99/328/head -> origin/gh/janeyx99/328/head 2025-12-04T09:16:47.4523212Z * [new branch] gh/janeyx99/328/orig -> origin/gh/janeyx99/328/orig 2025-12-04T09:16:47.4525669Z * [new branch] gh/janeyx99/329/base -> origin/gh/janeyx99/329/base 2025-12-04T09:16:47.4527566Z * [new branch] gh/janeyx99/329/head -> origin/gh/janeyx99/329/head 2025-12-04T09:16:47.4529461Z * [new branch] gh/janeyx99/329/orig -> origin/gh/janeyx99/329/orig 2025-12-04T09:16:47.4532526Z * [new branch] gh/janeyx99/330/base -> origin/gh/janeyx99/330/base 2025-12-04T09:16:47.4534938Z * [new branch] gh/janeyx99/330/head -> origin/gh/janeyx99/330/head 2025-12-04T09:16:47.4536268Z * [new branch] gh/janeyx99/330/orig -> origin/gh/janeyx99/330/orig 2025-12-04T09:16:47.4539515Z * [new branch] gh/janeyx99/331/base -> origin/gh/janeyx99/331/base 2025-12-04T09:16:47.4540839Z * [new branch] gh/janeyx99/331/head -> origin/gh/janeyx99/331/head 2025-12-04T09:16:47.4542886Z * [new branch] gh/janeyx99/331/orig -> origin/gh/janeyx99/331/orig 2025-12-04T09:16:47.4545474Z * [new branch] gh/janeyx99/332/base -> origin/gh/janeyx99/332/base 2025-12-04T09:16:47.4547655Z * [new branch] gh/janeyx99/332/head -> origin/gh/janeyx99/332/head 2025-12-04T09:16:47.4548884Z * [new branch] gh/janeyx99/332/orig -> origin/gh/janeyx99/332/orig 2025-12-04T09:16:47.4551594Z * [new branch] gh/janeyx99/333/base -> origin/gh/janeyx99/333/base 2025-12-04T09:16:47.4553471Z * [new branch] gh/janeyx99/333/head -> origin/gh/janeyx99/333/head 2025-12-04T09:16:47.4555334Z * [new branch] gh/janeyx99/333/orig -> origin/gh/janeyx99/333/orig 2025-12-04T09:16:47.4557943Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-12-04T09:16:47.4559875Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-12-04T09:16:47.4561689Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-12-04T09:16:47.4564872Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-12-04T09:16:47.4566623Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-12-04T09:16:47.4569433Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-12-04T09:16:47.4571147Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-12-04T09:16:47.4572994Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-12-04T09:16:47.4576055Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-12-04T09:16:47.4577890Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-12-04T09:16:47.4579724Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-12-04T09:16:47.4582278Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-12-04T09:16:47.4584143Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-12-04T09:16:47.4585963Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-12-04T09:16:47.4588432Z * [new branch] gh/jansel/552/base -> origin/gh/jansel/552/base 2025-12-04T09:16:47.4590302Z * [new branch] gh/jansel/552/head -> origin/gh/jansel/552/head 2025-12-04T09:16:47.4592086Z * [new branch] gh/jansel/552/orig -> origin/gh/jansel/552/orig 2025-12-04T09:16:47.4594537Z * [new branch] gh/jansel/553/base -> origin/gh/jansel/553/base 2025-12-04T09:16:47.4596385Z * [new branch] gh/jansel/553/head -> origin/gh/jansel/553/head 2025-12-04T09:16:47.4598200Z * [new branch] gh/jansel/553/orig -> origin/gh/jansel/553/orig 2025-12-04T09:16:47.4600632Z * [new branch] gh/jansel/554/base -> origin/gh/jansel/554/base 2025-12-04T09:16:47.4602424Z * [new branch] gh/jansel/554/head -> origin/gh/jansel/554/head 2025-12-04T09:16:47.4604231Z * [new branch] gh/jansel/554/orig -> origin/gh/jansel/554/orig 2025-12-04T09:16:47.4606727Z * [new branch] gh/jansel/555/base -> origin/gh/jansel/555/base 2025-12-04T09:16:47.4608987Z * [new branch] gh/jansel/555/head -> origin/gh/jansel/555/head 2025-12-04T09:16:47.4610661Z * [new branch] gh/jansel/555/orig -> origin/gh/jansel/555/orig 2025-12-04T09:16:47.4613148Z * [new branch] gh/jansel/556/base -> origin/gh/jansel/556/base 2025-12-04T09:16:47.4615240Z * [new branch] gh/jansel/556/head -> origin/gh/jansel/556/head 2025-12-04T09:16:47.4617386Z * [new branch] gh/jansel/556/orig -> origin/gh/jansel/556/orig 2025-12-04T09:16:47.4621011Z * [new branch] gh/jansel/557/base -> origin/gh/jansel/557/base 2025-12-04T09:16:47.4623577Z * [new branch] gh/jansel/557/head -> origin/gh/jansel/557/head 2025-12-04T09:16:47.4626086Z * [new branch] gh/jansel/557/orig -> origin/gh/jansel/557/orig 2025-12-04T09:16:47.4629483Z * [new branch] gh/jansel/558/base -> origin/gh/jansel/558/base 2025-12-04T09:16:47.4631939Z * [new branch] gh/jansel/558/head -> origin/gh/jansel/558/head 2025-12-04T09:16:47.4634589Z * [new branch] gh/jansel/558/orig -> origin/gh/jansel/558/orig 2025-12-04T09:16:47.4638734Z * [new branch] gh/jansel/559/base -> origin/gh/jansel/559/base 2025-12-04T09:16:47.4641301Z * [new branch] gh/jansel/559/head -> origin/gh/jansel/559/head 2025-12-04T09:16:47.4643819Z * [new branch] gh/jansel/559/orig -> origin/gh/jansel/559/orig 2025-12-04T09:16:47.4647383Z * [new branch] gh/jansel/560/base -> origin/gh/jansel/560/base 2025-12-04T09:16:47.4649354Z * [new branch] gh/jansel/560/head -> origin/gh/jansel/560/head 2025-12-04T09:16:47.4652024Z * [new branch] gh/jansel/560/orig -> origin/gh/jansel/560/orig 2025-12-04T09:16:47.4654306Z * [new branch] gh/jansel/561/base -> origin/gh/jansel/561/base 2025-12-04T09:16:47.4656129Z * [new branch] gh/jansel/561/head -> origin/gh/jansel/561/head 2025-12-04T09:16:47.4657884Z * [new branch] gh/jansel/561/orig -> origin/gh/jansel/561/orig 2025-12-04T09:16:47.4660411Z * [new branch] gh/jansel/562/base -> origin/gh/jansel/562/base 2025-12-04T09:16:47.4662207Z * [new branch] gh/jansel/562/head -> origin/gh/jansel/562/head 2025-12-04T09:16:47.4664007Z * [new branch] gh/jansel/562/orig -> origin/gh/jansel/562/orig 2025-12-04T09:16:47.4666548Z * [new branch] gh/jansel/563/base -> origin/gh/jansel/563/base 2025-12-04T09:16:47.4668467Z * [new branch] gh/jansel/563/head -> origin/gh/jansel/563/head 2025-12-04T09:16:47.4670326Z * [new branch] gh/jansel/563/orig -> origin/gh/jansel/563/orig 2025-12-04T09:16:47.4673381Z * [new branch] gh/jansel/564/base -> origin/gh/jansel/564/base 2025-12-04T09:16:47.4675238Z * [new branch] gh/jansel/564/head -> origin/gh/jansel/564/head 2025-12-04T09:16:47.4677055Z * [new branch] gh/jansel/564/orig -> origin/gh/jansel/564/orig 2025-12-04T09:16:47.4679689Z * [new branch] gh/jansel/565/base -> origin/gh/jansel/565/base 2025-12-04T09:16:47.4681512Z * [new branch] gh/jansel/565/head -> origin/gh/jansel/565/head 2025-12-04T09:16:47.4683342Z * [new branch] gh/jansel/565/orig -> origin/gh/jansel/565/orig 2025-12-04T09:16:47.4685937Z * [new branch] gh/jansel/566/base -> origin/gh/jansel/566/base 2025-12-04T09:16:47.4687830Z * [new branch] gh/jansel/566/head -> origin/gh/jansel/566/head 2025-12-04T09:16:47.4689677Z * [new branch] gh/jansel/566/orig -> origin/gh/jansel/566/orig 2025-12-04T09:16:47.4692733Z * [new branch] gh/jansel/567/base -> origin/gh/jansel/567/base 2025-12-04T09:16:47.4694805Z * [new branch] gh/jansel/567/head -> origin/gh/jansel/567/head 2025-12-04T09:16:47.4696844Z * [new branch] gh/jansel/567/orig -> origin/gh/jansel/567/orig 2025-12-04T09:16:47.4699160Z * [new branch] gh/jansel/568/base -> origin/gh/jansel/568/base 2025-12-04T09:16:47.4701092Z * [new branch] gh/jansel/568/head -> origin/gh/jansel/568/head 2025-12-04T09:16:47.4702871Z * [new branch] gh/jansel/568/orig -> origin/gh/jansel/568/orig 2025-12-04T09:16:47.4705436Z * [new branch] gh/jansel/569/base -> origin/gh/jansel/569/base 2025-12-04T09:16:47.4707226Z * [new branch] gh/jansel/569/head -> origin/gh/jansel/569/head 2025-12-04T09:16:47.4709065Z * [new branch] gh/jansel/569/orig -> origin/gh/jansel/569/orig 2025-12-04T09:16:47.4711608Z * [new branch] gh/jansel/570/base -> origin/gh/jansel/570/base 2025-12-04T09:16:47.4713550Z * [new branch] gh/jansel/570/head -> origin/gh/jansel/570/head 2025-12-04T09:16:47.4715340Z * [new branch] gh/jansel/570/orig -> origin/gh/jansel/570/orig 2025-12-04T09:16:47.4717929Z * [new branch] gh/jansel/571/base -> origin/gh/jansel/571/base 2025-12-04T09:16:47.4719885Z * [new branch] gh/jansel/571/head -> origin/gh/jansel/571/head 2025-12-04T09:16:47.4721736Z * [new branch] gh/jansel/571/orig -> origin/gh/jansel/571/orig 2025-12-04T09:16:47.4724140Z * [new branch] gh/jansel/572/base -> origin/gh/jansel/572/base 2025-12-04T09:16:47.4725969Z * [new branch] gh/jansel/572/head -> origin/gh/jansel/572/head 2025-12-04T09:16:47.4727869Z * [new branch] gh/jansel/572/orig -> origin/gh/jansel/572/orig 2025-12-04T09:16:47.4730596Z * [new branch] gh/jansel/573/base -> origin/gh/jansel/573/base 2025-12-04T09:16:47.4732411Z * [new branch] gh/jansel/573/head -> origin/gh/jansel/573/head 2025-12-04T09:16:47.4734206Z * [new branch] gh/jansel/573/orig -> origin/gh/jansel/573/orig 2025-12-04T09:16:47.4737192Z * [new branch] gh/jansel/574/base -> origin/gh/jansel/574/base 2025-12-04T09:16:47.4739131Z * [new branch] gh/jansel/574/head -> origin/gh/jansel/574/head 2025-12-04T09:16:47.4741003Z * [new branch] gh/jansel/574/orig -> origin/gh/jansel/574/orig 2025-12-04T09:16:47.4743506Z * [new branch] gh/jansel/575/base -> origin/gh/jansel/575/base 2025-12-04T09:16:47.4745492Z * [new branch] gh/jansel/575/head -> origin/gh/jansel/575/head 2025-12-04T09:16:47.4747287Z * [new branch] gh/jansel/575/orig -> origin/gh/jansel/575/orig 2025-12-04T09:16:47.4749886Z * [new branch] gh/jansel/576/base -> origin/gh/jansel/576/base 2025-12-04T09:16:47.4751751Z * [new branch] gh/jansel/576/head -> origin/gh/jansel/576/head 2025-12-04T09:16:47.4753597Z * [new branch] gh/jansel/576/orig -> origin/gh/jansel/576/orig 2025-12-04T09:16:47.4756706Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-12-04T09:16:47.4758556Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-12-04T09:16:47.4760389Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-12-04T09:16:47.4762893Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-12-04T09:16:47.4764973Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-12-04T09:16:47.4766787Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-12-04T09:16:47.4770997Z * [new branch] gh/jerryzh168/1/base -> origin/gh/jerryzh168/1/base 2025-12-04T09:16:47.4772685Z * [new branch] gh/jerryzh168/1/head -> origin/gh/jerryzh168/1/head 2025-12-04T09:16:47.4774436Z * [new branch] gh/jerryzh168/1/orig -> origin/gh/jerryzh168/1/orig 2025-12-04T09:16:47.4777510Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-12-04T09:16:47.4779362Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-12-04T09:16:47.4781171Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-12-04T09:16:47.4783576Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-12-04T09:16:47.4785425Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-12-04T09:16:47.4787265Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-12-04T09:16:47.4789802Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-12-04T09:16:47.4791563Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-12-04T09:16:47.4793501Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-12-04T09:16:47.4795999Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-12-04T09:16:47.4797835Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-12-04T09:16:47.4799646Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-12-04T09:16:47.4802176Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-12-04T09:16:47.4803982Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-12-04T09:16:47.4805911Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-12-04T09:16:47.4808490Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-12-04T09:16:47.4810288Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-12-04T09:16:47.4812023Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-12-04T09:16:47.4814803Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-12-04T09:16:47.4817280Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-12-04T09:16:47.4819082Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-12-04T09:16:47.4821589Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-12-04T09:16:47.4823507Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-12-04T09:16:47.4825303Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-12-04T09:16:47.4828304Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-12-04T09:16:47.4830179Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-12-04T09:16:47.4832099Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-12-04T09:16:47.4834491Z * [new branch] gh/jiayisunx/85/base -> origin/gh/jiayisunx/85/base 2025-12-04T09:16:47.4836334Z * [new branch] gh/jiayisunx/85/head -> origin/gh/jiayisunx/85/head 2025-12-04T09:16:47.4840634Z * [new branch] gh/jiayisunx/85/orig -> origin/gh/jiayisunx/85/orig 2025-12-04T09:16:47.4843645Z * [new branch] gh/jiayisunx/86/base -> origin/gh/jiayisunx/86/base 2025-12-04T09:16:47.4845436Z * [new branch] gh/jiayisunx/86/head -> origin/gh/jiayisunx/86/head 2025-12-04T09:16:47.4847658Z * [new branch] gh/jiayisunx/86/orig -> origin/gh/jiayisunx/86/orig 2025-12-04T09:16:47.4850091Z * [new branch] gh/jiayisunx/87/base -> origin/gh/jiayisunx/87/base 2025-12-04T09:16:47.4851908Z * [new branch] gh/jiayisunx/87/head -> origin/gh/jiayisunx/87/head 2025-12-04T09:16:47.4854012Z * [new branch] gh/jiayisunx/87/orig -> origin/gh/jiayisunx/87/orig 2025-12-04T09:16:47.4856221Z * [new branch] gh/jiayisunx/88/base -> origin/gh/jiayisunx/88/base 2025-12-04T09:16:47.4858080Z * [new branch] gh/jiayisunx/88/head -> origin/gh/jiayisunx/88/head 2025-12-04T09:16:47.4859919Z * [new branch] gh/jiayisunx/88/orig -> origin/gh/jiayisunx/88/orig 2025-12-04T09:16:47.4862461Z * [new branch] gh/jiayisunx/89/base -> origin/gh/jiayisunx/89/base 2025-12-04T09:16:47.4864246Z * [new branch] gh/jiayisunx/89/head -> origin/gh/jiayisunx/89/head 2025-12-04T09:16:47.4866062Z * [new branch] gh/jiayisunx/89/orig -> origin/gh/jiayisunx/89/orig 2025-12-04T09:16:47.4876575Z * [new branch] gh/jiayisunx/90/base -> origin/gh/jiayisunx/90/base 2025-12-04T09:16:47.4877080Z * [new branch] gh/jiayisunx/90/head -> origin/gh/jiayisunx/90/head 2025-12-04T09:16:47.4877330Z * [new branch] gh/jiayisunx/90/orig -> origin/gh/jiayisunx/90/orig 2025-12-04T09:16:47.4877566Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-12-04T09:16:47.4877787Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-12-04T09:16:47.4879915Z * [new branch] gh/jturney/1/base -> origin/gh/jturney/1/base 2025-12-04T09:16:47.4881678Z * [new branch] gh/jturney/1/head -> origin/gh/jturney/1/head 2025-12-04T09:16:47.4883489Z * [new branch] gh/jturney/1/orig -> origin/gh/jturney/1/orig 2025-12-04T09:16:47.4886078Z * [new branch] gh/jturney/2/base -> origin/gh/jturney/2/base 2025-12-04T09:16:47.4887979Z * [new branch] gh/jturney/2/head -> origin/gh/jturney/2/head 2025-12-04T09:16:47.4889916Z * [new branch] gh/jturney/2/orig -> origin/gh/jturney/2/orig 2025-12-04T09:16:47.4893752Z * [new branch] gh/karthickai/10/base -> origin/gh/karthickai/10/base 2025-12-04T09:16:47.4895649Z * [new branch] gh/karthickai/10/head -> origin/gh/karthickai/10/head 2025-12-04T09:16:47.4897489Z * [new branch] gh/karthickai/10/orig -> origin/gh/karthickai/10/orig 2025-12-04T09:16:47.4899991Z * [new branch] gh/karthickai/11/base -> origin/gh/karthickai/11/base 2025-12-04T09:16:47.4901893Z * [new branch] gh/karthickai/11/head -> origin/gh/karthickai/11/head 2025-12-04T09:16:47.4903741Z * [new branch] gh/karthickai/11/orig -> origin/gh/karthickai/11/orig 2025-12-04T09:16:47.4907097Z * [new branch] gh/karthickai/12/base -> origin/gh/karthickai/12/base 2025-12-04T09:16:47.4909006Z * [new branch] gh/karthickai/12/head -> origin/gh/karthickai/12/head 2025-12-04T09:16:47.4910854Z * [new branch] gh/karthickai/12/orig -> origin/gh/karthickai/12/orig 2025-12-04T09:16:47.4913370Z * [new branch] gh/karthickai/13/base -> origin/gh/karthickai/13/base 2025-12-04T09:16:47.4915438Z * [new branch] gh/karthickai/13/head -> origin/gh/karthickai/13/head 2025-12-04T09:16:47.4917308Z * [new branch] gh/karthickai/13/orig -> origin/gh/karthickai/13/orig 2025-12-04T09:16:47.4920058Z * [new branch] gh/karthickai/14/base -> origin/gh/karthickai/14/base 2025-12-04T09:16:47.4922098Z * [new branch] gh/karthickai/14/head -> origin/gh/karthickai/14/head 2025-12-04T09:16:47.4924089Z * [new branch] gh/karthickai/14/orig -> origin/gh/karthickai/14/orig 2025-12-04T09:16:47.4926757Z * [new branch] gh/karthickai/15/base -> origin/gh/karthickai/15/base 2025-12-04T09:16:47.4928869Z * [new branch] gh/karthickai/15/head -> origin/gh/karthickai/15/head 2025-12-04T09:16:47.4930682Z * [new branch] gh/karthickai/15/orig -> origin/gh/karthickai/15/orig 2025-12-04T09:16:47.4933125Z * [new branch] gh/karthickai/16/base -> origin/gh/karthickai/16/base 2025-12-04T09:16:47.4934989Z * [new branch] gh/karthickai/16/head -> origin/gh/karthickai/16/head 2025-12-04T09:16:47.4937316Z * [new branch] gh/karthickai/16/orig -> origin/gh/karthickai/16/orig 2025-12-04T09:16:47.4939712Z * [new branch] gh/karthickai/17/base -> origin/gh/karthickai/17/base 2025-12-04T09:16:47.4941505Z * [new branch] gh/karthickai/17/head -> origin/gh/karthickai/17/head 2025-12-04T09:16:47.4943301Z * [new branch] gh/karthickai/17/orig -> origin/gh/karthickai/17/orig 2025-12-04T09:16:47.4946060Z * [new branch] gh/karthickai/18/base -> origin/gh/karthickai/18/base 2025-12-04T09:16:47.4948048Z * [new branch] gh/karthickai/18/head -> origin/gh/karthickai/18/head 2025-12-04T09:16:47.4949984Z * [new branch] gh/karthickai/18/orig -> origin/gh/karthickai/18/orig 2025-12-04T09:16:47.4952564Z * [new branch] gh/karthickai/19/base -> origin/gh/karthickai/19/base 2025-12-04T09:16:47.4954411Z * [new branch] gh/karthickai/19/head -> origin/gh/karthickai/19/head 2025-12-04T09:16:47.4956200Z * [new branch] gh/karthickai/19/orig -> origin/gh/karthickai/19/orig 2025-12-04T09:16:47.4959616Z * [new branch] gh/karthickai/20/base -> origin/gh/karthickai/20/base 2025-12-04T09:16:47.4962193Z * [new branch] gh/karthickai/20/head -> origin/gh/karthickai/20/head 2025-12-04T09:16:47.4964093Z * [new branch] gh/karthickai/20/orig -> origin/gh/karthickai/20/orig 2025-12-04T09:16:47.4966614Z * [new branch] gh/karthickai/21/base -> origin/gh/karthickai/21/base 2025-12-04T09:16:47.4968839Z * [new branch] gh/karthickai/21/head -> origin/gh/karthickai/21/head 2025-12-04T09:16:47.4970691Z * [new branch] gh/karthickai/21/orig -> origin/gh/karthickai/21/orig 2025-12-04T09:16:47.4973371Z * [new branch] gh/karthickai/22/base -> origin/gh/karthickai/22/base 2025-12-04T09:16:47.4975157Z * [new branch] gh/karthickai/22/head -> origin/gh/karthickai/22/head 2025-12-04T09:16:47.4976974Z * [new branch] gh/karthickai/22/orig -> origin/gh/karthickai/22/orig 2025-12-04T09:16:47.4979703Z * [new branch] gh/karthickai/23/base -> origin/gh/karthickai/23/base 2025-12-04T09:16:47.4981740Z * [new branch] gh/karthickai/23/head -> origin/gh/karthickai/23/head 2025-12-04T09:16:47.4983535Z * [new branch] gh/karthickai/23/orig -> origin/gh/karthickai/23/orig 2025-12-04T09:16:47.4986061Z * [new branch] gh/karthickai/24/base -> origin/gh/karthickai/24/base 2025-12-04T09:16:47.4987909Z * [new branch] gh/karthickai/24/head -> origin/gh/karthickai/24/head 2025-12-04T09:16:47.4989798Z * [new branch] gh/karthickai/24/orig -> origin/gh/karthickai/24/orig 2025-12-04T09:16:47.4992854Z * [new branch] gh/karthickai/25/base -> origin/gh/karthickai/25/base 2025-12-04T09:16:47.4994794Z * [new branch] gh/karthickai/25/head -> origin/gh/karthickai/25/head 2025-12-04T09:16:47.4996616Z * [new branch] gh/karthickai/25/orig -> origin/gh/karthickai/25/orig 2025-12-04T09:16:47.4999107Z * [new branch] gh/karthickai/26/base -> origin/gh/karthickai/26/base 2025-12-04T09:16:47.5001175Z * [new branch] gh/karthickai/26/head -> origin/gh/karthickai/26/head 2025-12-04T09:16:47.5002871Z * [new branch] gh/karthickai/26/orig -> origin/gh/karthickai/26/orig 2025-12-04T09:16:47.5006630Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-12-04T09:16:47.5009294Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-12-04T09:16:47.5011102Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-12-04T09:16:47.5014177Z * [new branch] gh/krocki/1/base -> origin/gh/krocki/1/base 2025-12-04T09:16:47.5016008Z * [new branch] gh/krocki/1/head -> origin/gh/krocki/1/head 2025-12-04T09:16:47.5017871Z * [new branch] gh/krocki/1/orig -> origin/gh/krocki/1/orig 2025-12-04T09:16:47.5020380Z * [new branch] gh/krocki/2/base -> origin/gh/krocki/2/base 2025-12-04T09:16:47.5022211Z * [new branch] gh/krocki/2/head -> origin/gh/krocki/2/head 2025-12-04T09:16:47.5023953Z * [new branch] gh/krocki/2/orig -> origin/gh/krocki/2/orig 2025-12-04T09:16:47.5027180Z * [new branch] gh/kurtamohler/60/base -> origin/gh/kurtamohler/60/base 2025-12-04T09:16:47.5029038Z * [new branch] gh/kurtamohler/60/head -> origin/gh/kurtamohler/60/head 2025-12-04T09:16:47.5030869Z * [new branch] gh/kurtamohler/60/orig -> origin/gh/kurtamohler/60/orig 2025-12-04T09:16:47.5033390Z * [new branch] gh/kurtamohler/61/base -> origin/gh/kurtamohler/61/base 2025-12-04T09:16:47.5035137Z * [new branch] gh/kurtamohler/61/head -> origin/gh/kurtamohler/61/head 2025-12-04T09:16:47.5037181Z * [new branch] gh/kurtamohler/61/orig -> origin/gh/kurtamohler/61/orig 2025-12-04T09:16:47.5039612Z * [new branch] gh/kurtamohler/62/base -> origin/gh/kurtamohler/62/base 2025-12-04T09:16:47.5041442Z * [new branch] gh/kurtamohler/62/head -> origin/gh/kurtamohler/62/head 2025-12-04T09:16:47.5043212Z * [new branch] gh/kurtamohler/62/orig -> origin/gh/kurtamohler/62/orig 2025-12-04T09:16:47.5045730Z * [new branch] gh/kurtamohler/63/base -> origin/gh/kurtamohler/63/base 2025-12-04T09:16:47.5047584Z * [new branch] gh/kurtamohler/63/head -> origin/gh/kurtamohler/63/head 2025-12-04T09:16:47.5050116Z * [new branch] gh/kurtamohler/63/orig -> origin/gh/kurtamohler/63/orig 2025-12-04T09:16:47.5052673Z * [new branch] gh/kurtamohler/64/base -> origin/gh/kurtamohler/64/base 2025-12-04T09:16:47.5054474Z * [new branch] gh/kurtamohler/64/head -> origin/gh/kurtamohler/64/head 2025-12-04T09:16:47.5056323Z * [new branch] gh/kurtamohler/64/orig -> origin/gh/kurtamohler/64/orig 2025-12-04T09:16:47.5058862Z * [new branch] gh/kurtamohler/65/base -> origin/gh/kurtamohler/65/base 2025-12-04T09:16:47.5060629Z * [new branch] gh/kurtamohler/65/head -> origin/gh/kurtamohler/65/head 2025-12-04T09:16:47.5062460Z * [new branch] gh/kurtamohler/65/orig -> origin/gh/kurtamohler/65/orig 2025-12-04T09:16:47.5064904Z * [new branch] gh/kurtamohler/66/base -> origin/gh/kurtamohler/66/base 2025-12-04T09:16:47.5066765Z * [new branch] gh/kurtamohler/66/head -> origin/gh/kurtamohler/66/head 2025-12-04T09:16:47.5068589Z * [new branch] gh/kurtamohler/66/orig -> origin/gh/kurtamohler/66/orig 2025-12-04T09:16:47.5071097Z * [new branch] gh/kurtamohler/67/base -> origin/gh/kurtamohler/67/base 2025-12-04T09:16:47.5072938Z * [new branch] gh/kurtamohler/67/head -> origin/gh/kurtamohler/67/head 2025-12-04T09:16:47.5075048Z * [new branch] gh/kurtamohler/67/orig -> origin/gh/kurtamohler/67/orig 2025-12-04T09:16:47.5078070Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-12-04T09:16:47.5080118Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-12-04T09:16:47.5081954Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-12-04T09:16:47.5084496Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-12-04T09:16:47.5086327Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-12-04T09:16:47.5089118Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-12-04T09:16:47.5091028Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-12-04T09:16:47.5092863Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-12-04T09:16:47.5095309Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-12-04T09:16:47.5097095Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-12-04T09:16:47.5099023Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-12-04T09:16:47.5101704Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-12-04T09:16:47.5103575Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-12-04T09:16:47.5106039Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-12-04T09:16:47.5107883Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-12-04T09:16:47.5109722Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-12-04T09:16:47.5112252Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-12-04T09:16:47.5114081Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-12-04T09:16:47.5115885Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-12-04T09:16:47.5118527Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-12-04T09:16:47.5120362Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-12-04T09:16:47.5122141Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-12-04T09:16:47.5124749Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-12-04T09:16:47.5126624Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-12-04T09:16:47.5128553Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-12-04T09:16:47.5131013Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-12-04T09:16:47.5132848Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-12-04T09:16:47.5134665Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-12-04T09:16:47.5137618Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-12-04T09:16:47.5139403Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-12-04T09:16:47.5141262Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-12-04T09:16:47.5143757Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-12-04T09:16:47.5145535Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-12-04T09:16:47.5147450Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-12-04T09:16:47.5150590Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-12-04T09:16:47.5151975Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-12-04T09:16:47.5153765Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-12-04T09:16:47.5156243Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-12-04T09:16:47.5158040Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-12-04T09:16:47.5159830Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-12-04T09:16:47.5162305Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-12-04T09:16:47.5164104Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-12-04T09:16:47.5165932Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-12-04T09:16:47.5168547Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-12-04T09:16:47.5170338Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-12-04T09:16:47.5172146Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-12-04T09:16:47.5175235Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-12-04T09:16:47.5177224Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-12-04T09:16:47.5179019Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-12-04T09:16:47.5182469Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-12-04T09:16:47.5184445Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-12-04T09:16:47.5186282Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-12-04T09:16:47.5188841Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-12-04T09:16:47.5190673Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-12-04T09:16:47.5192483Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-12-04T09:16:47.5194995Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-12-04T09:16:47.5196927Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-12-04T09:16:47.5198793Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-12-04T09:16:47.5201476Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-12-04T09:16:47.5203375Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-12-04T09:16:47.5205237Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-12-04T09:16:47.5208035Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-12-04T09:16:47.5209944Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-12-04T09:16:47.5211770Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-12-04T09:16:47.5214401Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-12-04T09:16:47.5216386Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-12-04T09:16:47.5218663Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-12-04T09:16:47.5221321Z * [new branch] gh/kwen2501/275/base -> origin/gh/kwen2501/275/base 2025-12-04T09:16:47.5223322Z * [new branch] gh/kwen2501/275/head -> origin/gh/kwen2501/275/head 2025-12-04T09:16:47.5225261Z * [new branch] gh/kwen2501/275/orig -> origin/gh/kwen2501/275/orig 2025-12-04T09:16:47.5227902Z * [new branch] gh/kwen2501/276/base -> origin/gh/kwen2501/276/base 2025-12-04T09:16:47.5229706Z * [new branch] gh/kwen2501/276/head -> origin/gh/kwen2501/276/head 2025-12-04T09:16:47.5231511Z * [new branch] gh/kwen2501/276/orig -> origin/gh/kwen2501/276/orig 2025-12-04T09:16:47.5234093Z * [new branch] gh/kwen2501/277/base -> origin/gh/kwen2501/277/base 2025-12-04T09:16:47.5236057Z * [new branch] gh/kwen2501/277/head -> origin/gh/kwen2501/277/head 2025-12-04T09:16:47.5240329Z * [new branch] gh/kwen2501/277/orig -> origin/gh/kwen2501/277/orig 2025-12-04T09:16:47.5242863Z * [new branch] gh/kwen2501/278/base -> origin/gh/kwen2501/278/base 2025-12-04T09:16:47.5244694Z * [new branch] gh/kwen2501/278/head -> origin/gh/kwen2501/278/head 2025-12-04T09:16:47.5246535Z * [new branch] gh/kwen2501/278/orig -> origin/gh/kwen2501/278/orig 2025-12-04T09:16:47.5249412Z * [new branch] gh/kwen2501/279/base -> origin/gh/kwen2501/279/base 2025-12-04T09:16:47.5251325Z * [new branch] gh/kwen2501/279/head -> origin/gh/kwen2501/279/head 2025-12-04T09:16:47.5253200Z * [new branch] gh/kwen2501/279/orig -> origin/gh/kwen2501/279/orig 2025-12-04T09:16:47.5255869Z * [new branch] gh/kwen2501/280/base -> origin/gh/kwen2501/280/base 2025-12-04T09:16:47.5257709Z * [new branch] gh/kwen2501/280/head -> origin/gh/kwen2501/280/head 2025-12-04T09:16:47.5259567Z * [new branch] gh/kwen2501/280/orig -> origin/gh/kwen2501/280/orig 2025-12-04T09:16:47.5262139Z * [new branch] gh/kwen2501/281/base -> origin/gh/kwen2501/281/base 2025-12-04T09:16:47.5263991Z * [new branch] gh/kwen2501/281/head -> origin/gh/kwen2501/281/head 2025-12-04T09:16:47.5265900Z * [new branch] gh/kwen2501/281/orig -> origin/gh/kwen2501/281/orig 2025-12-04T09:16:47.5268619Z * [new branch] gh/kwen2501/282/base -> origin/gh/kwen2501/282/base 2025-12-04T09:16:47.5270546Z * [new branch] gh/kwen2501/282/head -> origin/gh/kwen2501/282/head 2025-12-04T09:16:47.5272311Z * [new branch] gh/kwen2501/282/orig -> origin/gh/kwen2501/282/orig 2025-12-04T09:16:47.5274903Z * [new branch] gh/kwen2501/283/base -> origin/gh/kwen2501/283/base 2025-12-04T09:16:47.5276803Z * [new branch] gh/kwen2501/283/head -> origin/gh/kwen2501/283/head 2025-12-04T09:16:47.5278632Z * [new branch] gh/kwen2501/283/orig -> origin/gh/kwen2501/283/orig 2025-12-04T09:16:47.5281315Z * [new branch] gh/kwen2501/284/base -> origin/gh/kwen2501/284/base 2025-12-04T09:16:47.5283172Z * [new branch] gh/kwen2501/284/head -> origin/gh/kwen2501/284/head 2025-12-04T09:16:47.5285050Z * [new branch] gh/kwen2501/284/orig -> origin/gh/kwen2501/284/orig 2025-12-04T09:16:47.5287642Z * [new branch] gh/kwen2501/285/base -> origin/gh/kwen2501/285/base 2025-12-04T09:16:47.5289547Z * [new branch] gh/kwen2501/285/head -> origin/gh/kwen2501/285/head 2025-12-04T09:16:47.5291316Z * [new branch] gh/kwen2501/285/orig -> origin/gh/kwen2501/285/orig 2025-12-04T09:16:47.5293910Z * [new branch] gh/kwen2501/286/base -> origin/gh/kwen2501/286/base 2025-12-04T09:16:47.5295779Z * [new branch] gh/kwen2501/286/head -> origin/gh/kwen2501/286/head 2025-12-04T09:16:47.5297619Z * [new branch] gh/kwen2501/286/orig -> origin/gh/kwen2501/286/orig 2025-12-04T09:16:47.5300067Z * [new branch] gh/kwen2501/287/base -> origin/gh/kwen2501/287/base 2025-12-04T09:16:47.5302004Z * [new branch] gh/kwen2501/287/head -> origin/gh/kwen2501/287/head 2025-12-04T09:16:47.5303688Z * [new branch] gh/kwen2501/287/orig -> origin/gh/kwen2501/287/orig 2025-12-04T09:16:47.5306401Z * [new branch] gh/kwen2501/288/base -> origin/gh/kwen2501/288/base 2025-12-04T09:16:47.5308419Z * [new branch] gh/kwen2501/288/head -> origin/gh/kwen2501/288/head 2025-12-04T09:16:47.5311028Z * [new branch] gh/kwen2501/288/orig -> origin/gh/kwen2501/288/orig 2025-12-04T09:16:47.5314111Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-12-04T09:16:47.5315902Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-12-04T09:16:47.5317699Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-12-04T09:16:47.5320768Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-12-04T09:16:47.5322519Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-12-04T09:16:47.5324313Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-12-04T09:16:47.5326901Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-12-04T09:16:47.5329628Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-12-04T09:16:47.5331993Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-12-04T09:16:47.5333807Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-12-04T09:16:47.5336447Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-12-04T09:16:47.5338452Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-12-04T09:16:47.5341073Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-12-04T09:16:47.5342861Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-12-04T09:16:47.5344626Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-12-04T09:16:47.5347479Z * [new branch] gh/laithsakka/316/base -> origin/gh/laithsakka/316/base 2025-12-04T09:16:47.5349185Z * [new branch] gh/laithsakka/316/head -> origin/gh/laithsakka/316/head 2025-12-04T09:16:47.5350975Z * [new branch] gh/laithsakka/316/orig -> origin/gh/laithsakka/316/orig 2025-12-04T09:16:47.5353637Z * [new branch] gh/laithsakka/317/base -> origin/gh/laithsakka/317/base 2025-12-04T09:16:47.5355418Z * [new branch] gh/laithsakka/317/head -> origin/gh/laithsakka/317/head 2025-12-04T09:16:47.5357228Z * [new branch] gh/laithsakka/317/orig -> origin/gh/laithsakka/317/orig 2025-12-04T09:16:47.5360199Z * [new branch] gh/laithsakka/319/base -> origin/gh/laithsakka/319/base 2025-12-04T09:16:47.5362016Z * [new branch] gh/laithsakka/319/head -> origin/gh/laithsakka/319/head 2025-12-04T09:16:47.5363809Z * [new branch] gh/laithsakka/319/orig -> origin/gh/laithsakka/319/orig 2025-12-04T09:16:47.5366240Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-12-04T09:16:47.5368014Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-12-04T09:16:47.5370719Z * [new branch] gh/laithsakka/320/base -> origin/gh/laithsakka/320/base 2025-12-04T09:16:47.5372533Z * [new branch] gh/laithsakka/320/head -> origin/gh/laithsakka/320/head 2025-12-04T09:16:47.5374343Z * [new branch] gh/laithsakka/320/orig -> origin/gh/laithsakka/320/orig 2025-12-04T09:16:47.5376877Z * [new branch] gh/laithsakka/321/base -> origin/gh/laithsakka/321/base 2025-12-04T09:16:47.5379069Z * [new branch] gh/laithsakka/321/head -> origin/gh/laithsakka/321/head 2025-12-04T09:16:47.5380632Z * [new branch] gh/laithsakka/321/orig -> origin/gh/laithsakka/321/orig 2025-12-04T09:16:47.5383345Z * [new branch] gh/laithsakka/322/base -> origin/gh/laithsakka/322/base 2025-12-04T09:16:47.5385227Z * [new branch] gh/laithsakka/322/head -> origin/gh/laithsakka/322/head 2025-12-04T09:16:47.5386986Z * [new branch] gh/laithsakka/322/orig -> origin/gh/laithsakka/322/orig 2025-12-04T09:16:47.5389633Z * [new branch] gh/laithsakka/323/base -> origin/gh/laithsakka/323/base 2025-12-04T09:16:47.5391601Z * [new branch] gh/laithsakka/323/head -> origin/gh/laithsakka/323/head 2025-12-04T09:16:47.5393423Z * [new branch] gh/laithsakka/323/orig -> origin/gh/laithsakka/323/orig 2025-12-04T09:16:47.5395969Z * [new branch] gh/laithsakka/324/base -> origin/gh/laithsakka/324/base 2025-12-04T09:16:47.5398233Z * [new branch] gh/laithsakka/324/head -> origin/gh/laithsakka/324/head 2025-12-04T09:16:47.5400128Z * [new branch] gh/laithsakka/324/orig -> origin/gh/laithsakka/324/orig 2025-12-04T09:16:47.5402664Z * [new branch] gh/laithsakka/325/base -> origin/gh/laithsakka/325/base 2025-12-04T09:16:47.5404718Z * [new branch] gh/laithsakka/325/head -> origin/gh/laithsakka/325/head 2025-12-04T09:16:47.5406409Z * [new branch] gh/laithsakka/325/orig -> origin/gh/laithsakka/325/orig 2025-12-04T09:16:47.5409629Z * [new branch] gh/laithsakka/326/base -> origin/gh/laithsakka/326/base 2025-12-04T09:16:47.5411473Z * [new branch] gh/laithsakka/326/head -> origin/gh/laithsakka/326/head 2025-12-04T09:16:47.5413376Z * [new branch] gh/laithsakka/326/orig -> origin/gh/laithsakka/326/orig 2025-12-04T09:16:47.5415928Z * [new branch] gh/laithsakka/327/base -> origin/gh/laithsakka/327/base 2025-12-04T09:16:47.5417784Z * [new branch] gh/laithsakka/327/head -> origin/gh/laithsakka/327/head 2025-12-04T09:16:47.5419633Z * [new branch] gh/laithsakka/327/orig -> origin/gh/laithsakka/327/orig 2025-12-04T09:16:47.5422286Z * [new branch] gh/laithsakka/328/base -> origin/gh/laithsakka/328/base 2025-12-04T09:16:47.5423986Z * [new branch] gh/laithsakka/328/head -> origin/gh/laithsakka/328/head 2025-12-04T09:16:47.5425933Z * [new branch] gh/laithsakka/328/orig -> origin/gh/laithsakka/328/orig 2025-12-04T09:16:47.5429216Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-12-04T09:16:47.5431030Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-12-04T09:16:47.5433156Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-12-04T09:16:47.5438173Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-12-04T09:16:47.5439988Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-12-04T09:16:47.5442920Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-12-04T09:16:47.5444731Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-12-04T09:16:47.5446553Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-12-04T09:16:47.5449221Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-12-04T09:16:47.5451092Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-12-04T09:16:47.5452884Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-12-04T09:16:47.5455452Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-12-04T09:16:47.5457450Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-12-04T09:16:47.5459139Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-12-04T09:16:47.5462044Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-12-04T09:16:47.5464605Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-12-04T09:16:47.5466451Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-12-04T09:16:47.5468218Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-12-04T09:16:47.5470769Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-12-04T09:16:47.5472605Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-12-04T09:16:47.5474932Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-12-04T09:16:47.5477513Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-12-04T09:16:47.5479329Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-12-04T09:16:47.5481781Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-12-04T09:16:47.5483512Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-12-04T09:16:47.5485374Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-12-04T09:16:47.5487872Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-12-04T09:16:47.5490014Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-12-04T09:16:47.5491997Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-12-04T09:16:47.5494937Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-12-04T09:16:47.5496505Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-12-04T09:16:47.5498370Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-12-04T09:16:47.5500871Z * [new branch] gh/malfet/565/base -> origin/gh/malfet/565/base 2025-12-04T09:16:47.5502787Z * [new branch] gh/malfet/565/head -> origin/gh/malfet/565/head 2025-12-04T09:16:47.5504643Z * [new branch] gh/malfet/565/orig -> origin/gh/malfet/565/orig 2025-12-04T09:16:47.5507237Z * [new branch] gh/malfet/575/base -> origin/gh/malfet/575/base 2025-12-04T09:16:47.5509004Z * [new branch] gh/malfet/575/head -> origin/gh/malfet/575/head 2025-12-04T09:16:47.5510824Z * [new branch] gh/malfet/575/orig -> origin/gh/malfet/575/orig 2025-12-04T09:16:47.5513258Z * [new branch] gh/malfet/580/base -> origin/gh/malfet/580/base 2025-12-04T09:16:47.5515141Z * [new branch] gh/malfet/580/head -> origin/gh/malfet/580/head 2025-12-04T09:16:47.5516940Z * [new branch] gh/malfet/580/orig -> origin/gh/malfet/580/orig 2025-12-04T09:16:47.5519388Z * [new branch] gh/malfet/581/base -> origin/gh/malfet/581/base 2025-12-04T09:16:47.5521235Z * [new branch] gh/malfet/581/head -> origin/gh/malfet/581/head 2025-12-04T09:16:47.5523051Z * [new branch] gh/malfet/581/orig -> origin/gh/malfet/581/orig 2025-12-04T09:16:47.5525467Z * [new branch] gh/malfet/583/base -> origin/gh/malfet/583/base 2025-12-04T09:16:47.5527488Z * [new branch] gh/malfet/583/head -> origin/gh/malfet/583/head 2025-12-04T09:16:47.5529996Z * [new branch] gh/malfet/583/orig -> origin/gh/malfet/583/orig 2025-12-04T09:16:47.5532455Z * [new branch] gh/malfet/586/base -> origin/gh/malfet/586/base 2025-12-04T09:16:47.5534430Z * [new branch] gh/malfet/586/head -> origin/gh/malfet/586/head 2025-12-04T09:16:47.5536849Z * [new branch] gh/malfet/586/orig -> origin/gh/malfet/586/orig 2025-12-04T09:16:47.5539473Z * [new branch] gh/malfet/587/base -> origin/gh/malfet/587/base 2025-12-04T09:16:47.5541325Z * [new branch] gh/malfet/587/head -> origin/gh/malfet/587/head 2025-12-04T09:16:47.5543084Z * [new branch] gh/malfet/587/orig -> origin/gh/malfet/587/orig 2025-12-04T09:16:47.5545544Z * [new branch] gh/malfet/588/base -> origin/gh/malfet/588/base 2025-12-04T09:16:47.5547474Z * [new branch] gh/malfet/588/head -> origin/gh/malfet/588/head 2025-12-04T09:16:47.5549428Z * [new branch] gh/malfet/588/orig -> origin/gh/malfet/588/orig 2025-12-04T09:16:47.5551990Z * [new branch] gh/malfet/589/base -> origin/gh/malfet/589/base 2025-12-04T09:16:47.5553943Z * [new branch] gh/malfet/589/head -> origin/gh/malfet/589/head 2025-12-04T09:16:47.5555769Z * [new branch] gh/malfet/589/orig -> origin/gh/malfet/589/orig 2025-12-04T09:16:47.5558230Z * [new branch] gh/malfet/590/base -> origin/gh/malfet/590/base 2025-12-04T09:16:47.5560236Z * [new branch] gh/malfet/590/head -> origin/gh/malfet/590/head 2025-12-04T09:16:47.5561824Z * [new branch] gh/malfet/590/orig -> origin/gh/malfet/590/orig 2025-12-04T09:16:47.5565431Z * [new branch] gh/malfet/591/base -> origin/gh/malfet/591/base 2025-12-04T09:16:47.5567209Z * [new branch] gh/malfet/591/head -> origin/gh/malfet/591/head 2025-12-04T09:16:47.5569319Z * [new branch] gh/malfet/591/orig -> origin/gh/malfet/591/orig 2025-12-04T09:16:47.5571771Z * [new branch] gh/malfet/592/base -> origin/gh/malfet/592/base 2025-12-04T09:16:47.5573729Z * [new branch] gh/malfet/592/head -> origin/gh/malfet/592/head 2025-12-04T09:16:47.5575440Z * [new branch] gh/malfet/592/orig -> origin/gh/malfet/592/orig 2025-12-04T09:16:47.5578132Z * [new branch] gh/malfet/593/base -> origin/gh/malfet/593/base 2025-12-04T09:16:47.5579965Z * [new branch] gh/malfet/593/head -> origin/gh/malfet/593/head 2025-12-04T09:16:47.5581980Z * [new branch] gh/malfet/593/orig -> origin/gh/malfet/593/orig 2025-12-04T09:16:47.5584413Z * [new branch] gh/malfet/594/base -> origin/gh/malfet/594/base 2025-12-04T09:16:47.5586231Z * [new branch] gh/malfet/594/head -> origin/gh/malfet/594/head 2025-12-04T09:16:47.5588175Z * [new branch] gh/malfet/594/orig -> origin/gh/malfet/594/orig 2025-12-04T09:16:47.5590489Z * [new branch] gh/malfet/595/base -> origin/gh/malfet/595/base 2025-12-04T09:16:47.5592356Z * [new branch] gh/malfet/595/head -> origin/gh/malfet/595/head 2025-12-04T09:16:47.5594197Z * [new branch] gh/malfet/595/orig -> origin/gh/malfet/595/orig 2025-12-04T09:16:47.5596752Z * [new branch] gh/malfet/596/base -> origin/gh/malfet/596/base 2025-12-04T09:16:47.5598581Z * [new branch] gh/malfet/596/head -> origin/gh/malfet/596/head 2025-12-04T09:16:47.5600495Z * [new branch] gh/malfet/596/orig -> origin/gh/malfet/596/orig 2025-12-04T09:16:47.5602956Z * [new branch] gh/malfet/597/base -> origin/gh/malfet/597/base 2025-12-04T09:16:47.5604845Z * [new branch] gh/malfet/597/head -> origin/gh/malfet/597/head 2025-12-04T09:16:47.5606659Z * [new branch] gh/malfet/597/orig -> origin/gh/malfet/597/orig 2025-12-04T09:16:47.5609390Z * [new branch] gh/malfet/598/base -> origin/gh/malfet/598/base 2025-12-04T09:16:47.5611305Z * [new branch] gh/malfet/598/head -> origin/gh/malfet/598/head 2025-12-04T09:16:47.5613017Z * [new branch] gh/malfet/598/orig -> origin/gh/malfet/598/orig 2025-12-04T09:16:47.5615477Z * [new branch] gh/malfet/599/base -> origin/gh/malfet/599/base 2025-12-04T09:16:47.5617386Z * [new branch] gh/malfet/599/head -> origin/gh/malfet/599/head 2025-12-04T09:16:47.5619256Z * [new branch] gh/malfet/599/orig -> origin/gh/malfet/599/orig 2025-12-04T09:16:47.5621792Z * [new branch] gh/malfet/600/base -> origin/gh/malfet/600/base 2025-12-04T09:16:47.5623742Z * [new branch] gh/malfet/600/head -> origin/gh/malfet/600/head 2025-12-04T09:16:47.5625567Z * [new branch] gh/malfet/600/orig -> origin/gh/malfet/600/orig 2025-12-04T09:16:47.5628649Z * [new branch] gh/malfet/601/base -> origin/gh/malfet/601/base 2025-12-04T09:16:47.5630676Z * [new branch] gh/malfet/601/head -> origin/gh/malfet/601/head 2025-12-04T09:16:47.5632983Z * [new branch] gh/malfet/601/orig -> origin/gh/malfet/601/orig 2025-12-04T09:16:47.5635710Z * [new branch] gh/malfet/602/base -> origin/gh/malfet/602/base 2025-12-04T09:16:47.5639994Z * [new branch] gh/malfet/602/head -> origin/gh/malfet/602/head 2025-12-04T09:16:47.5641744Z * [new branch] gh/malfet/602/orig -> origin/gh/malfet/602/orig 2025-12-04T09:16:47.5644265Z * [new branch] gh/malfet/603/base -> origin/gh/malfet/603/base 2025-12-04T09:16:47.5645967Z * [new branch] gh/malfet/603/head -> origin/gh/malfet/603/head 2025-12-04T09:16:47.5647811Z * [new branch] gh/malfet/603/orig -> origin/gh/malfet/603/orig 2025-12-04T09:16:47.5650528Z * [new branch] gh/malfet/604/base -> origin/gh/malfet/604/base 2025-12-04T09:16:47.5652350Z * [new branch] gh/malfet/604/head -> origin/gh/malfet/604/head 2025-12-04T09:16:47.5654169Z * [new branch] gh/malfet/604/orig -> origin/gh/malfet/604/orig 2025-12-04T09:16:47.5656737Z * [new branch] gh/malfet/605/base -> origin/gh/malfet/605/base 2025-12-04T09:16:47.5658642Z * [new branch] gh/malfet/605/head -> origin/gh/malfet/605/head 2025-12-04T09:16:47.5660480Z * [new branch] gh/malfet/605/orig -> origin/gh/malfet/605/orig 2025-12-04T09:16:47.5663535Z * [new branch] gh/malfet/606/base -> origin/gh/malfet/606/base 2025-12-04T09:16:47.5665496Z * [new branch] gh/malfet/606/head -> origin/gh/malfet/606/head 2025-12-04T09:16:47.5667371Z * [new branch] gh/malfet/606/orig -> origin/gh/malfet/606/orig 2025-12-04T09:16:47.5669980Z * [new branch] gh/malfet/607/base -> origin/gh/malfet/607/base 2025-12-04T09:16:47.5671827Z * [new branch] gh/malfet/607/head -> origin/gh/malfet/607/head 2025-12-04T09:16:47.5673662Z * [new branch] gh/malfet/607/orig -> origin/gh/malfet/607/orig 2025-12-04T09:16:47.5676223Z * [new branch] gh/malfet/608/base -> origin/gh/malfet/608/base 2025-12-04T09:16:47.5678069Z * [new branch] gh/malfet/608/head -> origin/gh/malfet/608/head 2025-12-04T09:16:47.5679975Z * [new branch] gh/malfet/608/orig -> origin/gh/malfet/608/orig 2025-12-04T09:16:47.5682526Z * [new branch] gh/malfet/609/base -> origin/gh/malfet/609/base 2025-12-04T09:16:47.5684488Z * [new branch] gh/malfet/609/head -> origin/gh/malfet/609/head 2025-12-04T09:16:47.5686373Z * [new branch] gh/malfet/609/orig -> origin/gh/malfet/609/orig 2025-12-04T09:16:47.5689230Z * [new branch] gh/malfet/610/base -> origin/gh/malfet/610/base 2025-12-04T09:16:47.5690967Z * [new branch] gh/malfet/610/head -> origin/gh/malfet/610/head 2025-12-04T09:16:47.5692753Z * [new branch] gh/malfet/610/orig -> origin/gh/malfet/610/orig 2025-12-04T09:16:47.5695303Z * [new branch] gh/malfet/611/base -> origin/gh/malfet/611/base 2025-12-04T09:16:47.5697149Z * [new branch] gh/malfet/611/head -> origin/gh/malfet/611/head 2025-12-04T09:16:47.5698946Z * [new branch] gh/malfet/611/orig -> origin/gh/malfet/611/orig 2025-12-04T09:16:47.5701495Z * [new branch] gh/malfet/612/base -> origin/gh/malfet/612/base 2025-12-04T09:16:47.5703292Z * [new branch] gh/malfet/612/head -> origin/gh/malfet/612/head 2025-12-04T09:16:47.5705142Z * [new branch] gh/malfet/612/orig -> origin/gh/malfet/612/orig 2025-12-04T09:16:47.5707791Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-12-04T09:16:47.5709705Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-12-04T09:16:47.5712861Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-12-04T09:16:47.5714619Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-12-04T09:16:47.5716489Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-12-04T09:16:47.5719622Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-12-04T09:16:47.5722662Z * [new branch] gh/masnesral/1/base -> origin/gh/masnesral/1/base 2025-12-04T09:16:47.5724518Z * [new branch] gh/masnesral/1/head -> origin/gh/masnesral/1/head 2025-12-04T09:16:47.5726352Z * [new branch] gh/masnesral/1/orig -> origin/gh/masnesral/1/orig 2025-12-04T09:16:47.5729818Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-12-04T09:16:47.5731632Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-12-04T09:16:47.5733921Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-12-04T09:16:47.5735905Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-12-04T09:16:47.5738485Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-12-04T09:16:47.5740428Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-12-04T09:16:47.5742729Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-12-04T09:16:47.5744501Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-12-04T09:16:47.5746908Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-12-04T09:16:47.5748791Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-12-04T09:16:47.5751146Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-12-04T09:16:47.5752976Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-12-04T09:16:47.5755575Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-12-04T09:16:47.5757403Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-12-04T09:16:47.5760465Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-12-04T09:16:47.5762426Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-12-04T09:16:47.5764679Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-12-04T09:16:47.5766682Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-12-04T09:16:47.5769491Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-12-04T09:16:47.5771276Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-12-04T09:16:47.5773697Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-12-04T09:16:47.5775587Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-12-04T09:16:47.5777998Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-12-04T09:16:47.5779810Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-12-04T09:16:47.5782308Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-12-04T09:16:47.5784164Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-12-04T09:16:47.5785967Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-12-04T09:16:47.5789250Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-12-04T09:16:47.5791223Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-12-04T09:16:47.5792847Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-12-04T09:16:47.5795575Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-12-04T09:16:47.5797359Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-12-04T09:16:47.5799236Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-12-04T09:16:47.5801880Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-12-04T09:16:47.5803598Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-12-04T09:16:47.5805450Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-12-04T09:16:47.5808265Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-12-04T09:16:47.5810139Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-12-04T09:16:47.5811987Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-12-04T09:16:47.5814581Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-12-04T09:16:47.5816401Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-12-04T09:16:47.5818323Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-12-04T09:16:47.5820751Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-12-04T09:16:47.5822549Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-12-04T09:16:47.5824417Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-12-04T09:16:47.5827498Z * [new branch] gh/mikaylagawarecki/351/base -> origin/gh/mikaylagawarecki/351/base 2025-12-04T09:16:47.5829404Z * [new branch] gh/mikaylagawarecki/351/head -> origin/gh/mikaylagawarecki/351/head 2025-12-04T09:16:47.5831155Z * [new branch] gh/mikaylagawarecki/351/orig -> origin/gh/mikaylagawarecki/351/orig 2025-12-04T09:16:47.5833876Z * [new branch] gh/mikaylagawarecki/352/base -> origin/gh/mikaylagawarecki/352/base 2025-12-04T09:16:47.5835976Z * [new branch] gh/mikaylagawarecki/352/head -> origin/gh/mikaylagawarecki/352/head 2025-12-04T09:16:47.5838060Z * [new branch] gh/mikaylagawarecki/352/orig -> origin/gh/mikaylagawarecki/352/orig 2025-12-04T09:16:47.5840672Z * [new branch] gh/mikaylagawarecki/353/base -> origin/gh/mikaylagawarecki/353/base 2025-12-04T09:16:47.5842802Z * [new branch] gh/mikaylagawarecki/353/head -> origin/gh/mikaylagawarecki/353/head 2025-12-04T09:16:47.5844638Z * [new branch] gh/mikaylagawarecki/353/orig -> origin/gh/mikaylagawarecki/353/orig 2025-12-04T09:16:47.5846967Z * [new branch] gh/mikaylagawarecki/354/base -> origin/gh/mikaylagawarecki/354/base 2025-12-04T09:16:47.5848984Z * [new branch] gh/mikaylagawarecki/354/head -> origin/gh/mikaylagawarecki/354/head 2025-12-04T09:16:47.5850923Z * [new branch] gh/mikaylagawarecki/354/orig -> origin/gh/mikaylagawarecki/354/orig 2025-12-04T09:16:47.5854488Z * [new branch] gh/mikaylagawarecki/356/base -> origin/gh/mikaylagawarecki/356/base 2025-12-04T09:16:47.5856384Z * [new branch] gh/mikaylagawarecki/356/head -> origin/gh/mikaylagawarecki/356/head 2025-12-04T09:16:47.5858225Z * [new branch] gh/mikaylagawarecki/356/orig -> origin/gh/mikaylagawarecki/356/orig 2025-12-04T09:16:47.5860660Z * [new branch] gh/mikaylagawarecki/357/base -> origin/gh/mikaylagawarecki/357/base 2025-12-04T09:16:47.5862517Z * [new branch] gh/mikaylagawarecki/357/head -> origin/gh/mikaylagawarecki/357/head 2025-12-04T09:16:47.5864312Z * [new branch] gh/mikaylagawarecki/357/orig -> origin/gh/mikaylagawarecki/357/orig 2025-12-04T09:16:47.5867068Z * [new branch] gh/mikaylagawarecki/359/base -> origin/gh/mikaylagawarecki/359/base 2025-12-04T09:16:47.5869015Z * [new branch] gh/mikaylagawarecki/359/head -> origin/gh/mikaylagawarecki/359/head 2025-12-04T09:16:47.5870895Z * [new branch] gh/mikaylagawarecki/359/orig -> origin/gh/mikaylagawarecki/359/orig 2025-12-04T09:16:47.5873498Z * [new branch] gh/mikaylagawarecki/360/base -> origin/gh/mikaylagawarecki/360/base 2025-12-04T09:16:47.5875359Z * [new branch] gh/mikaylagawarecki/360/head -> origin/gh/mikaylagawarecki/360/head 2025-12-04T09:16:47.5877162Z * [new branch] gh/mikaylagawarecki/360/orig -> origin/gh/mikaylagawarecki/360/orig 2025-12-04T09:16:47.5879915Z * [new branch] gh/mikaylagawarecki/361/base -> origin/gh/mikaylagawarecki/361/base 2025-12-04T09:16:47.5882016Z * [new branch] gh/mikaylagawarecki/361/head -> origin/gh/mikaylagawarecki/361/head 2025-12-04T09:16:47.5883675Z * [new branch] gh/mikaylagawarecki/361/orig -> origin/gh/mikaylagawarecki/361/orig 2025-12-04T09:16:47.5886292Z * [new branch] gh/mikaylagawarecki/362/base -> origin/gh/mikaylagawarecki/362/base 2025-12-04T09:16:47.5888463Z * [new branch] gh/mikaylagawarecki/362/head -> origin/gh/mikaylagawarecki/362/head 2025-12-04T09:16:47.5890297Z * [new branch] gh/mikaylagawarecki/362/orig -> origin/gh/mikaylagawarecki/362/orig 2025-12-04T09:16:47.5893195Z * [new branch] gh/mikaylagawarecki/363/base -> origin/gh/mikaylagawarecki/363/base 2025-12-04T09:16:47.5895161Z * [new branch] gh/mikaylagawarecki/363/head -> origin/gh/mikaylagawarecki/363/head 2025-12-04T09:16:47.5896993Z * [new branch] gh/mikaylagawarecki/363/orig -> origin/gh/mikaylagawarecki/363/orig 2025-12-04T09:16:47.5900008Z * [new branch] gh/mikaylagawarecki/364/base -> origin/gh/mikaylagawarecki/364/base 2025-12-04T09:16:47.5901873Z * [new branch] gh/mikaylagawarecki/364/head -> origin/gh/mikaylagawarecki/364/head 2025-12-04T09:16:47.5903683Z * [new branch] gh/mikaylagawarecki/364/orig -> origin/gh/mikaylagawarecki/364/orig 2025-12-04T09:16:47.5906514Z * [new branch] gh/mikaylagawarecki/365/base -> origin/gh/mikaylagawarecki/365/base 2025-12-04T09:16:47.5908508Z * [new branch] gh/mikaylagawarecki/365/head -> origin/gh/mikaylagawarecki/365/head 2025-12-04T09:16:47.5910410Z * [new branch] gh/mikaylagawarecki/365/orig -> origin/gh/mikaylagawarecki/365/orig 2025-12-04T09:16:47.5913129Z * [new branch] gh/mikaylagawarecki/366/base -> origin/gh/mikaylagawarecki/366/base 2025-12-04T09:16:47.5914710Z * [new branch] gh/mikaylagawarecki/366/head -> origin/gh/mikaylagawarecki/366/head 2025-12-04T09:16:47.5916564Z * [new branch] gh/mikaylagawarecki/366/orig -> origin/gh/mikaylagawarecki/366/orig 2025-12-04T09:16:47.5919177Z * [new branch] gh/mikaylagawarecki/367/base -> origin/gh/mikaylagawarecki/367/base 2025-12-04T09:16:47.5920955Z * [new branch] gh/mikaylagawarecki/367/head -> origin/gh/mikaylagawarecki/367/head 2025-12-04T09:16:47.5922789Z * [new branch] gh/mikaylagawarecki/367/orig -> origin/gh/mikaylagawarecki/367/orig 2025-12-04T09:16:47.5925360Z * [new branch] gh/mikaylagawarecki/368/base -> origin/gh/mikaylagawarecki/368/base 2025-12-04T09:16:47.5927241Z * [new branch] gh/mikaylagawarecki/368/head -> origin/gh/mikaylagawarecki/368/head 2025-12-04T09:16:47.5929333Z * [new branch] gh/mikaylagawarecki/368/orig -> origin/gh/mikaylagawarecki/368/orig 2025-12-04T09:16:47.5932038Z * [new branch] gh/mikaylagawarecki/369/base -> origin/gh/mikaylagawarecki/369/base 2025-12-04T09:16:47.5933942Z * [new branch] gh/mikaylagawarecki/369/head -> origin/gh/mikaylagawarecki/369/head 2025-12-04T09:16:47.5936183Z * [new branch] gh/mikaylagawarecki/369/orig -> origin/gh/mikaylagawarecki/369/orig 2025-12-04T09:16:47.5938984Z * [new branch] gh/mikaylagawarecki/370/base -> origin/gh/mikaylagawarecki/370/base 2025-12-04T09:16:47.5940828Z * [new branch] gh/mikaylagawarecki/370/head -> origin/gh/mikaylagawarecki/370/head 2025-12-04T09:16:47.5942651Z * [new branch] gh/mikaylagawarecki/370/orig -> origin/gh/mikaylagawarecki/370/orig 2025-12-04T09:16:47.5945225Z * [new branch] gh/mikaylagawarecki/371/base -> origin/gh/mikaylagawarecki/371/base 2025-12-04T09:16:47.5947088Z * [new branch] gh/mikaylagawarecki/371/head -> origin/gh/mikaylagawarecki/371/head 2025-12-04T09:16:47.5948925Z * [new branch] gh/mikaylagawarecki/371/orig -> origin/gh/mikaylagawarecki/371/orig 2025-12-04T09:16:47.5951627Z * [new branch] gh/mikaylagawarecki/372/base -> origin/gh/mikaylagawarecki/372/base 2025-12-04T09:16:47.5953431Z * [new branch] gh/mikaylagawarecki/372/head -> origin/gh/mikaylagawarecki/372/head 2025-12-04T09:16:47.5955181Z * [new branch] gh/mikaylagawarecki/372/orig -> origin/gh/mikaylagawarecki/372/orig 2025-12-04T09:16:47.5957794Z * [new branch] gh/mikaylagawarecki/373/base -> origin/gh/mikaylagawarecki/373/base 2025-12-04T09:16:47.5959638Z * [new branch] gh/mikaylagawarecki/373/head -> origin/gh/mikaylagawarecki/373/head 2025-12-04T09:16:47.5961431Z * [new branch] gh/mikaylagawarecki/373/orig -> origin/gh/mikaylagawarecki/373/orig 2025-12-04T09:16:47.5963986Z * [new branch] gh/mikaylagawarecki/374/base -> origin/gh/mikaylagawarecki/374/base 2025-12-04T09:16:47.5965855Z * [new branch] gh/mikaylagawarecki/374/head -> origin/gh/mikaylagawarecki/374/head 2025-12-04T09:16:47.5967723Z * [new branch] gh/mikaylagawarecki/374/orig -> origin/gh/mikaylagawarecki/374/orig 2025-12-04T09:16:47.5970464Z * [new branch] gh/mikaylagawarecki/375/base -> origin/gh/mikaylagawarecki/375/base 2025-12-04T09:16:47.5972330Z * [new branch] gh/mikaylagawarecki/375/head -> origin/gh/mikaylagawarecki/375/head 2025-12-04T09:16:47.5974154Z * [new branch] gh/mikaylagawarecki/375/orig -> origin/gh/mikaylagawarecki/375/orig 2025-12-04T09:16:47.5976841Z * [new branch] gh/mikaylagawarecki/376/base -> origin/gh/mikaylagawarecki/376/base 2025-12-04T09:16:47.5978826Z * [new branch] gh/mikaylagawarecki/376/head -> origin/gh/mikaylagawarecki/376/head 2025-12-04T09:16:47.5980661Z * [new branch] gh/mikaylagawarecki/376/orig -> origin/gh/mikaylagawarecki/376/orig 2025-12-04T09:16:47.5983268Z * [new branch] gh/mikaylagawarecki/377/base -> origin/gh/mikaylagawarecki/377/base 2025-12-04T09:16:47.5985218Z * [new branch] gh/mikaylagawarecki/377/head -> origin/gh/mikaylagawarecki/377/head 2025-12-04T09:16:47.5987081Z * [new branch] gh/mikaylagawarecki/377/orig -> origin/gh/mikaylagawarecki/377/orig 2025-12-04T09:16:47.5989656Z * [new branch] gh/mikaylagawarecki/378/base -> origin/gh/mikaylagawarecki/378/base 2025-12-04T09:16:47.5991556Z * [new branch] gh/mikaylagawarecki/378/head -> origin/gh/mikaylagawarecki/378/head 2025-12-04T09:16:47.5993464Z * [new branch] gh/mikaylagawarecki/378/orig -> origin/gh/mikaylagawarecki/378/orig 2025-12-04T09:16:47.5995954Z * [new branch] gh/mikaylagawarecki/379/base -> origin/gh/mikaylagawarecki/379/base 2025-12-04T09:16:47.5997842Z * [new branch] gh/mikaylagawarecki/379/head -> origin/gh/mikaylagawarecki/379/head 2025-12-04T09:16:47.5999640Z * [new branch] gh/mikaylagawarecki/379/orig -> origin/gh/mikaylagawarecki/379/orig 2025-12-04T09:16:47.6002298Z * [new branch] gh/mikaylagawarecki/380/base -> origin/gh/mikaylagawarecki/380/base 2025-12-04T09:16:47.6003959Z * [new branch] gh/mikaylagawarecki/380/head -> origin/gh/mikaylagawarecki/380/head 2025-12-04T09:16:47.6005792Z * [new branch] gh/mikaylagawarecki/380/orig -> origin/gh/mikaylagawarecki/380/orig 2025-12-04T09:16:47.6008531Z * [new branch] gh/mikaylagawarecki/381/base -> origin/gh/mikaylagawarecki/381/base 2025-12-04T09:16:47.6010369Z * [new branch] gh/mikaylagawarecki/381/head -> origin/gh/mikaylagawarecki/381/head 2025-12-04T09:16:47.6012175Z * [new branch] gh/mikaylagawarecki/381/orig -> origin/gh/mikaylagawarecki/381/orig 2025-12-04T09:16:47.6014559Z * [new branch] gh/mikaylagawarecki/382/base -> origin/gh/mikaylagawarecki/382/base 2025-12-04T09:16:47.6016894Z * [new branch] gh/mikaylagawarecki/382/head -> origin/gh/mikaylagawarecki/382/head 2025-12-04T09:16:47.6018696Z * [new branch] gh/mikaylagawarecki/382/orig -> origin/gh/mikaylagawarecki/382/orig 2025-12-04T09:16:47.6021378Z * [new branch] gh/mikaylagawarecki/383/base -> origin/gh/mikaylagawarecki/383/base 2025-12-04T09:16:47.6023268Z * [new branch] gh/mikaylagawarecki/383/head -> origin/gh/mikaylagawarecki/383/head 2025-12-04T09:16:47.6025512Z * [new branch] gh/mikaylagawarecki/383/orig -> origin/gh/mikaylagawarecki/383/orig 2025-12-04T09:16:47.6028037Z * [new branch] gh/mikaylagawarecki/384/base -> origin/gh/mikaylagawarecki/384/base 2025-12-04T09:16:47.6029946Z * [new branch] gh/mikaylagawarecki/384/head -> origin/gh/mikaylagawarecki/384/head 2025-12-04T09:16:47.6031731Z * [new branch] gh/mikaylagawarecki/384/orig -> origin/gh/mikaylagawarecki/384/orig 2025-12-04T09:16:47.6034938Z * [new branch] gh/mikaylagawarecki/385/base -> origin/gh/mikaylagawarecki/385/base 2025-12-04T09:16:47.6039318Z * [new branch] gh/mikaylagawarecki/385/head -> origin/gh/mikaylagawarecki/385/head 2025-12-04T09:16:47.6041123Z * [new branch] gh/mikaylagawarecki/385/orig -> origin/gh/mikaylagawarecki/385/orig 2025-12-04T09:16:47.6043780Z * [new branch] gh/mikaylagawarecki/386/base -> origin/gh/mikaylagawarecki/386/base 2025-12-04T09:16:47.6045590Z * [new branch] gh/mikaylagawarecki/386/head -> origin/gh/mikaylagawarecki/386/head 2025-12-04T09:16:47.6047505Z * [new branch] gh/mikaylagawarecki/386/orig -> origin/gh/mikaylagawarecki/386/orig 2025-12-04T09:16:47.6050404Z * [new branch] gh/mikaylagawarecki/387/base -> origin/gh/mikaylagawarecki/387/base 2025-12-04T09:16:47.6052111Z * [new branch] gh/mikaylagawarecki/387/head -> origin/gh/mikaylagawarecki/387/head 2025-12-04T09:16:47.6053971Z * [new branch] gh/mikaylagawarecki/387/orig -> origin/gh/mikaylagawarecki/387/orig 2025-12-04T09:16:47.6056399Z * [new branch] gh/mikaylagawarecki/388/base -> origin/gh/mikaylagawarecki/388/base 2025-12-04T09:16:47.6058217Z * [new branch] gh/mikaylagawarecki/388/head -> origin/gh/mikaylagawarecki/388/head 2025-12-04T09:16:47.6060182Z * [new branch] gh/mikaylagawarecki/388/orig -> origin/gh/mikaylagawarecki/388/orig 2025-12-04T09:16:47.6062884Z * [new branch] gh/mikaylagawarecki/389/base -> origin/gh/mikaylagawarecki/389/base 2025-12-04T09:16:47.6064667Z * [new branch] gh/mikaylagawarecki/389/head -> origin/gh/mikaylagawarecki/389/head 2025-12-04T09:16:47.6066522Z * [new branch] gh/mikaylagawarecki/389/orig -> origin/gh/mikaylagawarecki/389/orig 2025-12-04T09:16:47.6069141Z * [new branch] gh/mikaylagawarecki/390/base -> origin/gh/mikaylagawarecki/390/base 2025-12-04T09:16:47.6070920Z * [new branch] gh/mikaylagawarecki/390/head -> origin/gh/mikaylagawarecki/390/head 2025-12-04T09:16:47.6072701Z * [new branch] gh/mikaylagawarecki/390/orig -> origin/gh/mikaylagawarecki/390/orig 2025-12-04T09:16:47.6075404Z * [new branch] gh/mikaylagawarecki/391/base -> origin/gh/mikaylagawarecki/391/base 2025-12-04T09:16:47.6077364Z * [new branch] gh/mikaylagawarecki/391/head -> origin/gh/mikaylagawarecki/391/head 2025-12-04T09:16:47.6079339Z * [new branch] gh/mikaylagawarecki/391/orig -> origin/gh/mikaylagawarecki/391/orig 2025-12-04T09:16:47.6081822Z * [new branch] gh/mikaylagawarecki/392/base -> origin/gh/mikaylagawarecki/392/base 2025-12-04T09:16:47.6083644Z * [new branch] gh/mikaylagawarecki/392/head -> origin/gh/mikaylagawarecki/392/head 2025-12-04T09:16:47.6085681Z * [new branch] gh/mikaylagawarecki/392/orig -> origin/gh/mikaylagawarecki/392/orig 2025-12-04T09:16:47.6088814Z * [new branch] gh/mlazos/41/base -> origin/gh/mlazos/41/base 2025-12-04T09:16:47.6090585Z * [new branch] gh/mlazos/41/head -> origin/gh/mlazos/41/head 2025-12-04T09:16:47.6092386Z * [new branch] gh/mlazos/41/orig -> origin/gh/mlazos/41/orig 2025-12-04T09:16:47.6094979Z * [new branch] gh/mlazos/42/base -> origin/gh/mlazos/42/base 2025-12-04T09:16:47.6097159Z * [new branch] gh/mlazos/42/head -> origin/gh/mlazos/42/head 2025-12-04T09:16:47.6099008Z * [new branch] gh/mlazos/42/orig -> origin/gh/mlazos/42/orig 2025-12-04T09:16:47.6102046Z * [new branch] gh/mlazos/43/base -> origin/gh/mlazos/43/base 2025-12-04T09:16:47.6104020Z * [new branch] gh/mlazos/43/head -> origin/gh/mlazos/43/head 2025-12-04T09:16:47.6105769Z * [new branch] gh/mlazos/43/orig -> origin/gh/mlazos/43/orig 2025-12-04T09:16:47.6108130Z * [new branch] gh/mlazos/44/base -> origin/gh/mlazos/44/base 2025-12-04T09:16:47.6110003Z * [new branch] gh/mlazos/44/head -> origin/gh/mlazos/44/head 2025-12-04T09:16:47.6111960Z * [new branch] gh/mlazos/44/orig -> origin/gh/mlazos/44/orig 2025-12-04T09:16:47.6114347Z * [new branch] gh/mlazos/47/base -> origin/gh/mlazos/47/base 2025-12-04T09:16:47.6116199Z * [new branch] gh/mlazos/47/head -> origin/gh/mlazos/47/head 2025-12-04T09:16:47.6117985Z * [new branch] gh/mlazos/47/orig -> origin/gh/mlazos/47/orig 2025-12-04T09:16:47.6120392Z * [new branch] gh/mlazos/48/base -> origin/gh/mlazos/48/base 2025-12-04T09:16:47.6122367Z * [new branch] gh/mlazos/48/head -> origin/gh/mlazos/48/head 2025-12-04T09:16:47.6124412Z * [new branch] gh/mlazos/48/orig -> origin/gh/mlazos/48/orig 2025-12-04T09:16:47.6126520Z * [new branch] gh/mlazos/49/base -> origin/gh/mlazos/49/base 2025-12-04T09:16:47.6128511Z * [new branch] gh/mlazos/49/head -> origin/gh/mlazos/49/head 2025-12-04T09:16:47.6130511Z * [new branch] gh/mlazos/49/orig -> origin/gh/mlazos/49/orig 2025-12-04T09:16:47.6132773Z * [new branch] gh/mlazos/50/base -> origin/gh/mlazos/50/base 2025-12-04T09:16:47.6134547Z * [new branch] gh/mlazos/50/head -> origin/gh/mlazos/50/head 2025-12-04T09:16:47.6136856Z * [new branch] gh/mlazos/50/orig -> origin/gh/mlazos/50/orig 2025-12-04T09:16:47.6139378Z * [new branch] gh/mlazos/51/base -> origin/gh/mlazos/51/base 2025-12-04T09:16:47.6141211Z * [new branch] gh/mlazos/51/head -> origin/gh/mlazos/51/head 2025-12-04T09:16:47.6143000Z * [new branch] gh/mlazos/51/orig -> origin/gh/mlazos/51/orig 2025-12-04T09:16:47.6145545Z * [new branch] gh/mlazos/52/base -> origin/gh/mlazos/52/base 2025-12-04T09:16:47.6147290Z * [new branch] gh/mlazos/52/head -> origin/gh/mlazos/52/head 2025-12-04T09:16:47.6149206Z * [new branch] gh/mlazos/52/orig -> origin/gh/mlazos/52/orig 2025-12-04T09:16:47.6151649Z * [new branch] gh/mlazos/53/base -> origin/gh/mlazos/53/base 2025-12-04T09:16:47.6153574Z * [new branch] gh/mlazos/53/head -> origin/gh/mlazos/53/head 2025-12-04T09:16:47.6155286Z * [new branch] gh/mlazos/53/orig -> origin/gh/mlazos/53/orig 2025-12-04T09:16:47.6157781Z * [new branch] gh/mlazos/54/base -> origin/gh/mlazos/54/base 2025-12-04T09:16:47.6159588Z * [new branch] gh/mlazos/54/head -> origin/gh/mlazos/54/head 2025-12-04T09:16:47.6161564Z * [new branch] gh/mlazos/54/orig -> origin/gh/mlazos/54/orig 2025-12-04T09:16:47.6163937Z * [new branch] gh/mlazos/55/base -> origin/gh/mlazos/55/base 2025-12-04T09:16:47.6165675Z * [new branch] gh/mlazos/55/head -> origin/gh/mlazos/55/head 2025-12-04T09:16:47.6167564Z * [new branch] gh/mlazos/55/orig -> origin/gh/mlazos/55/orig 2025-12-04T09:16:47.6170265Z * [new branch] gh/mlazos/56/base -> origin/gh/mlazos/56/base 2025-12-04T09:16:47.6172033Z * [new branch] gh/mlazos/56/head -> origin/gh/mlazos/56/head 2025-12-04T09:16:47.6173860Z * [new branch] gh/mlazos/56/orig -> origin/gh/mlazos/56/orig 2025-12-04T09:16:47.6176369Z * [new branch] gh/mlazos/57/base -> origin/gh/mlazos/57/base 2025-12-04T09:16:47.6178160Z * [new branch] gh/mlazos/57/head -> origin/gh/mlazos/57/head 2025-12-04T09:16:47.6179944Z * [new branch] gh/mlazos/57/orig -> origin/gh/mlazos/57/orig 2025-12-04T09:16:47.6182755Z * [new branch] gh/mlazos/58/base -> origin/gh/mlazos/58/base 2025-12-04T09:16:47.6185079Z * [new branch] gh/mlazos/58/head -> origin/gh/mlazos/58/head 2025-12-04T09:16:47.6186895Z * [new branch] gh/mlazos/58/orig -> origin/gh/mlazos/58/orig 2025-12-04T09:16:47.6189433Z * [new branch] gh/mlazos/59/base -> origin/gh/mlazos/59/base 2025-12-04T09:16:47.6191202Z * [new branch] gh/mlazos/59/head -> origin/gh/mlazos/59/head 2025-12-04T09:16:47.6192923Z * [new branch] gh/mlazos/59/orig -> origin/gh/mlazos/59/orig 2025-12-04T09:16:47.6195502Z * [new branch] gh/mlazos/60/base -> origin/gh/mlazos/60/base 2025-12-04T09:16:47.6197462Z * [new branch] gh/mlazos/60/head -> origin/gh/mlazos/60/head 2025-12-04T09:16:47.6199129Z * [new branch] gh/mlazos/60/orig -> origin/gh/mlazos/60/orig 2025-12-04T09:16:47.6202100Z * [new branch] gh/mlazos/61/base -> origin/gh/mlazos/61/base 2025-12-04T09:16:47.6205699Z * [new branch] gh/mlazos/61/head -> origin/gh/mlazos/61/head 2025-12-04T09:16:47.6206239Z * [new branch] gh/mlazos/61/orig -> origin/gh/mlazos/61/orig 2025-12-04T09:16:47.6208852Z * [new branch] gh/mlazos/62/base -> origin/gh/mlazos/62/base 2025-12-04T09:16:47.6210603Z * [new branch] gh/mlazos/62/head -> origin/gh/mlazos/62/head 2025-12-04T09:16:47.6212479Z * [new branch] gh/mlazos/62/orig -> origin/gh/mlazos/62/orig 2025-12-04T09:16:47.6215122Z * [new branch] gh/mlazos/63/base -> origin/gh/mlazos/63/base 2025-12-04T09:16:47.6217039Z * [new branch] gh/mlazos/63/head -> origin/gh/mlazos/63/head 2025-12-04T09:16:47.6218873Z * [new branch] gh/mlazos/63/orig -> origin/gh/mlazos/63/orig 2025-12-04T09:16:47.6221269Z * [new branch] gh/mlazos/64/base -> origin/gh/mlazos/64/base 2025-12-04T09:16:47.6223635Z * [new branch] gh/mlazos/64/head -> origin/gh/mlazos/64/head 2025-12-04T09:16:47.6225469Z * [new branch] gh/mlazos/64/orig -> origin/gh/mlazos/64/orig 2025-12-04T09:16:47.6228037Z * [new branch] gh/mlazos/65/base -> origin/gh/mlazos/65/base 2025-12-04T09:16:47.6230410Z * [new branch] gh/mlazos/65/head -> origin/gh/mlazos/65/head 2025-12-04T09:16:47.6232235Z * [new branch] gh/mlazos/65/orig -> origin/gh/mlazos/65/orig 2025-12-04T09:16:47.6234785Z * [new branch] gh/mlazos/66/base -> origin/gh/mlazos/66/base 2025-12-04T09:16:47.6239114Z * [new branch] gh/mlazos/66/head -> origin/gh/mlazos/66/head 2025-12-04T09:16:47.6240906Z * [new branch] gh/mlazos/66/orig -> origin/gh/mlazos/66/orig 2025-12-04T09:16:47.6243409Z * [new branch] gh/mlazos/67/base -> origin/gh/mlazos/67/base 2025-12-04T09:16:47.6245245Z * [new branch] gh/mlazos/67/head -> origin/gh/mlazos/67/head 2025-12-04T09:16:47.6247145Z * [new branch] gh/mlazos/67/orig -> origin/gh/mlazos/67/orig 2025-12-04T09:16:47.6249865Z * [new branch] gh/mlazos/68/base -> origin/gh/mlazos/68/base 2025-12-04T09:16:47.6251748Z * [new branch] gh/mlazos/68/head -> origin/gh/mlazos/68/head 2025-12-04T09:16:47.6253579Z * [new branch] gh/mlazos/68/orig -> origin/gh/mlazos/68/orig 2025-12-04T09:16:47.6256185Z * [new branch] gh/mlazos/69/base -> origin/gh/mlazos/69/base 2025-12-04T09:16:47.6257953Z * [new branch] gh/mlazos/69/head -> origin/gh/mlazos/69/head 2025-12-04T09:16:47.6259759Z * [new branch] gh/mlazos/69/orig -> origin/gh/mlazos/69/orig 2025-12-04T09:16:47.6262301Z * [new branch] gh/mlazos/70/base -> origin/gh/mlazos/70/base 2025-12-04T09:16:47.6264113Z * [new branch] gh/mlazos/70/head -> origin/gh/mlazos/70/head 2025-12-04T09:16:47.6266071Z * [new branch] gh/mlazos/70/orig -> origin/gh/mlazos/70/orig 2025-12-04T09:16:47.6268657Z * [new branch] gh/mlazos/71/base -> origin/gh/mlazos/71/base 2025-12-04T09:16:47.6270470Z * [new branch] gh/mlazos/71/head -> origin/gh/mlazos/71/head 2025-12-04T09:16:47.6272268Z * [new branch] gh/mlazos/71/orig -> origin/gh/mlazos/71/orig 2025-12-04T09:16:47.6274769Z * [new branch] gh/mlazos/72/base -> origin/gh/mlazos/72/base 2025-12-04T09:16:47.6276844Z * [new branch] gh/mlazos/72/head -> origin/gh/mlazos/72/head 2025-12-04T09:16:47.6278575Z * [new branch] gh/mlazos/72/orig -> origin/gh/mlazos/72/orig 2025-12-04T09:16:47.6281116Z * [new branch] gh/mlazos/73/base -> origin/gh/mlazos/73/base 2025-12-04T09:16:47.6282899Z * [new branch] gh/mlazos/73/head -> origin/gh/mlazos/73/head 2025-12-04T09:16:47.6284711Z * [new branch] gh/mlazos/73/orig -> origin/gh/mlazos/73/orig 2025-12-04T09:16:47.6288480Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-12-04T09:16:47.6290559Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-12-04T09:16:47.6293570Z * [new branch] gh/muchulee8/73/base -> origin/gh/muchulee8/73/base 2025-12-04T09:16:47.6295618Z * [new branch] gh/muchulee8/73/head -> origin/gh/muchulee8/73/head 2025-12-04T09:16:47.6297526Z * [new branch] gh/muchulee8/73/orig -> origin/gh/muchulee8/73/orig 2025-12-04T09:16:47.6300529Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-12-04T09:16:47.6302503Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-12-04T09:16:47.6304435Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-12-04T09:16:47.6306856Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-12-04T09:16:47.6308808Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-12-04T09:16:47.6310719Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-12-04T09:16:47.6313140Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-12-04T09:16:47.6315018Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-12-04T09:16:47.6316855Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-12-04T09:16:47.6319327Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-12-04T09:16:47.6321118Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-12-04T09:16:47.6323202Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-12-04T09:16:47.6325528Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-12-04T09:16:47.6327367Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-12-04T09:16:47.6329605Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-12-04T09:16:47.6332203Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-12-04T09:16:47.6333980Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-12-04T09:16:47.6335685Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-12-04T09:16:47.6338679Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-12-04T09:16:47.6340590Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-12-04T09:16:47.6342327Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-12-04T09:16:47.6344619Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-12-04T09:16:47.6346665Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-12-04T09:16:47.6348524Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-12-04T09:16:47.6351088Z * [new branch] gh/naveenthangudu/9/base -> origin/gh/naveenthangudu/9/base 2025-12-04T09:16:47.6352883Z * [new branch] gh/naveenthangudu/9/head -> origin/gh/naveenthangudu/9/head 2025-12-04T09:16:47.6354758Z * [new branch] gh/naveenthangudu/9/orig -> origin/gh/naveenthangudu/9/orig 2025-12-04T09:16:47.6357641Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-12-04T09:16:47.6359557Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-12-04T09:16:47.6361277Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-12-04T09:16:47.6364146Z * [new branch] gh/nikitaved/10/base -> origin/gh/nikitaved/10/base 2025-12-04T09:16:47.6365975Z * [new branch] gh/nikitaved/10/head -> origin/gh/nikitaved/10/head 2025-12-04T09:16:47.6374394Z * [new branch] gh/nikitaved/10/orig -> origin/gh/nikitaved/10/orig 2025-12-04T09:16:47.6374868Z * [new branch] gh/nikitaved/11/base -> origin/gh/nikitaved/11/base 2025-12-04T09:16:47.6375107Z * [new branch] gh/nikitaved/11/head -> origin/gh/nikitaved/11/head 2025-12-04T09:16:47.6375327Z * [new branch] gh/nikitaved/11/orig -> origin/gh/nikitaved/11/orig 2025-12-04T09:16:47.6376136Z * [new branch] gh/nikitaved/12/base -> origin/gh/nikitaved/12/base 2025-12-04T09:16:47.6378151Z * [new branch] gh/nikitaved/12/head -> origin/gh/nikitaved/12/head 2025-12-04T09:16:47.6379968Z * [new branch] gh/nikitaved/12/orig -> origin/gh/nikitaved/12/orig 2025-12-04T09:16:47.6382489Z * [new branch] gh/nikitaved/13/base -> origin/gh/nikitaved/13/base 2025-12-04T09:16:47.6384254Z * [new branch] gh/nikitaved/13/head -> origin/gh/nikitaved/13/head 2025-12-04T09:16:47.6386094Z * [new branch] gh/nikitaved/13/orig -> origin/gh/nikitaved/13/orig 2025-12-04T09:16:47.6389008Z * [new branch] gh/nikitaved/14/base -> origin/gh/nikitaved/14/base 2025-12-04T09:16:47.6390698Z * [new branch] gh/nikitaved/14/head -> origin/gh/nikitaved/14/head 2025-12-04T09:16:47.6392530Z * [new branch] gh/nikitaved/14/orig -> origin/gh/nikitaved/14/orig 2025-12-04T09:16:47.6395140Z * [new branch] gh/nikitaved/15/base -> origin/gh/nikitaved/15/base 2025-12-04T09:16:47.6396779Z * [new branch] gh/nikitaved/15/head -> origin/gh/nikitaved/15/head 2025-12-04T09:16:47.6398686Z * [new branch] gh/nikitaved/15/orig -> origin/gh/nikitaved/15/orig 2025-12-04T09:16:47.6401104Z * [new branch] gh/nikitaved/16/base -> origin/gh/nikitaved/16/base 2025-12-04T09:16:47.6402936Z * [new branch] gh/nikitaved/16/head -> origin/gh/nikitaved/16/head 2025-12-04T09:16:47.6404735Z * [new branch] gh/nikitaved/16/orig -> origin/gh/nikitaved/16/orig 2025-12-04T09:16:47.6407295Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-12-04T09:16:47.6409278Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-12-04T09:16:47.6411103Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-12-04T09:16:47.6413659Z * [new branch] gh/nikitaved/4/base -> origin/gh/nikitaved/4/base 2025-12-04T09:16:47.6415422Z * [new branch] gh/nikitaved/4/head -> origin/gh/nikitaved/4/head 2025-12-04T09:16:47.6417278Z * [new branch] gh/nikitaved/4/orig -> origin/gh/nikitaved/4/orig 2025-12-04T09:16:47.6419865Z * [new branch] gh/nikitaved/5/base -> origin/gh/nikitaved/5/base 2025-12-04T09:16:47.6421668Z * [new branch] gh/nikitaved/5/head -> origin/gh/nikitaved/5/head 2025-12-04T09:16:47.6423629Z * [new branch] gh/nikitaved/5/orig -> origin/gh/nikitaved/5/orig 2025-12-04T09:16:47.6425904Z * [new branch] gh/nikitaved/6/base -> origin/gh/nikitaved/6/base 2025-12-04T09:16:47.6428379Z * [new branch] gh/nikitaved/6/head -> origin/gh/nikitaved/6/head 2025-12-04T09:16:47.6430061Z * [new branch] gh/nikitaved/6/orig -> origin/gh/nikitaved/6/orig 2025-12-04T09:16:47.6432541Z * [new branch] gh/nikitaved/8/base -> origin/gh/nikitaved/8/base 2025-12-04T09:16:47.6434345Z * [new branch] gh/nikitaved/8/head -> origin/gh/nikitaved/8/head 2025-12-04T09:16:47.6436425Z * [new branch] gh/nikitaved/8/orig -> origin/gh/nikitaved/8/orig 2025-12-04T09:16:47.6438936Z * [new branch] gh/nikitaved/9/base -> origin/gh/nikitaved/9/base 2025-12-04T09:16:47.6440669Z * [new branch] gh/nikitaved/9/head -> origin/gh/nikitaved/9/head 2025-12-04T09:16:47.6442488Z * [new branch] gh/nikitaved/9/orig -> origin/gh/nikitaved/9/orig 2025-12-04T09:16:47.6445485Z * [new branch] gh/oulgen/10/base -> origin/gh/oulgen/10/base 2025-12-04T09:16:47.6447633Z * [new branch] gh/oulgen/10/head -> origin/gh/oulgen/10/head 2025-12-04T09:16:47.6449315Z * [new branch] gh/oulgen/10/orig -> origin/gh/oulgen/10/orig 2025-12-04T09:16:47.6451688Z * [new branch] gh/oulgen/11/base -> origin/gh/oulgen/11/base 2025-12-04T09:16:47.6453487Z * [new branch] gh/oulgen/11/head -> origin/gh/oulgen/11/head 2025-12-04T09:16:47.6455277Z * [new branch] gh/oulgen/11/orig -> origin/gh/oulgen/11/orig 2025-12-04T09:16:47.6458255Z * [new branch] gh/oulgen/12/base -> origin/gh/oulgen/12/base 2025-12-04T09:16:47.6460045Z * [new branch] gh/oulgen/12/head -> origin/gh/oulgen/12/head 2025-12-04T09:16:47.6462041Z * [new branch] gh/oulgen/12/orig -> origin/gh/oulgen/12/orig 2025-12-04T09:16:47.6464432Z * [new branch] gh/oulgen/13/base -> origin/gh/oulgen/13/base 2025-12-04T09:16:47.6466293Z * [new branch] gh/oulgen/13/head -> origin/gh/oulgen/13/head 2025-12-04T09:16:47.6468082Z * [new branch] gh/oulgen/13/orig -> origin/gh/oulgen/13/orig 2025-12-04T09:16:47.6470616Z * [new branch] gh/oulgen/14/base -> origin/gh/oulgen/14/base 2025-12-04T09:16:47.6472447Z * [new branch] gh/oulgen/14/head -> origin/gh/oulgen/14/head 2025-12-04T09:16:47.6474563Z * [new branch] gh/oulgen/14/orig -> origin/gh/oulgen/14/orig 2025-12-04T09:16:47.6476808Z * [new branch] gh/oulgen/15/base -> origin/gh/oulgen/15/base 2025-12-04T09:16:47.6478636Z * [new branch] gh/oulgen/15/head -> origin/gh/oulgen/15/head 2025-12-04T09:16:47.6480485Z * [new branch] gh/oulgen/15/orig -> origin/gh/oulgen/15/orig 2025-12-04T09:16:47.6482834Z * [new branch] gh/oulgen/16/base -> origin/gh/oulgen/16/base 2025-12-04T09:16:47.6484618Z * [new branch] gh/oulgen/16/head -> origin/gh/oulgen/16/head 2025-12-04T09:16:47.6486481Z * [new branch] gh/oulgen/16/orig -> origin/gh/oulgen/16/orig 2025-12-04T09:16:47.6489157Z * [new branch] gh/oulgen/17/base -> origin/gh/oulgen/17/base 2025-12-04T09:16:47.6491047Z * [new branch] gh/oulgen/17/head -> origin/gh/oulgen/17/head 2025-12-04T09:16:47.6493335Z * [new branch] gh/oulgen/17/orig -> origin/gh/oulgen/17/orig 2025-12-04T09:16:47.6495401Z * [new branch] gh/oulgen/18/base -> origin/gh/oulgen/18/base 2025-12-04T09:16:47.6497184Z * [new branch] gh/oulgen/18/head -> origin/gh/oulgen/18/head 2025-12-04T09:16:47.6499235Z * [new branch] gh/oulgen/18/orig -> origin/gh/oulgen/18/orig 2025-12-04T09:16:47.6501580Z * [new branch] gh/oulgen/19/base -> origin/gh/oulgen/19/base 2025-12-04T09:16:47.6503366Z * [new branch] gh/oulgen/19/head -> origin/gh/oulgen/19/head 2025-12-04T09:16:47.6505266Z * [new branch] gh/oulgen/19/orig -> origin/gh/oulgen/19/orig 2025-12-04T09:16:47.6507744Z * [new branch] gh/oulgen/20/base -> origin/gh/oulgen/20/base 2025-12-04T09:16:47.6509586Z * [new branch] gh/oulgen/20/head -> origin/gh/oulgen/20/head 2025-12-04T09:16:47.6511511Z * [new branch] gh/oulgen/20/orig -> origin/gh/oulgen/20/orig 2025-12-04T09:16:47.6513819Z * [new branch] gh/oulgen/21/base -> origin/gh/oulgen/21/base 2025-12-04T09:16:47.6515695Z * [new branch] gh/oulgen/21/head -> origin/gh/oulgen/21/head 2025-12-04T09:16:47.6517734Z * [new branch] gh/oulgen/21/orig -> origin/gh/oulgen/21/orig 2025-12-04T09:16:47.6520081Z * [new branch] gh/oulgen/22/base -> origin/gh/oulgen/22/base 2025-12-04T09:16:47.6521883Z * [new branch] gh/oulgen/22/head -> origin/gh/oulgen/22/head 2025-12-04T09:16:47.6523652Z * [new branch] gh/oulgen/22/orig -> origin/gh/oulgen/22/orig 2025-12-04T09:16:47.6526145Z * [new branch] gh/oulgen/23/base -> origin/gh/oulgen/23/base 2025-12-04T09:16:47.6527984Z * [new branch] gh/oulgen/23/head -> origin/gh/oulgen/23/head 2025-12-04T09:16:47.6529847Z * [new branch] gh/oulgen/23/orig -> origin/gh/oulgen/23/orig 2025-12-04T09:16:47.6532224Z * [new branch] gh/oulgen/24/base -> origin/gh/oulgen/24/base 2025-12-04T09:16:47.6534151Z * [new branch] gh/oulgen/24/head -> origin/gh/oulgen/24/head 2025-12-04T09:16:47.6536149Z * [new branch] gh/oulgen/24/orig -> origin/gh/oulgen/24/orig 2025-12-04T09:16:47.6539043Z * [new branch] gh/oulgen/25/base -> origin/gh/oulgen/25/base 2025-12-04T09:16:47.6540850Z * [new branch] gh/oulgen/25/head -> origin/gh/oulgen/25/head 2025-12-04T09:16:47.6542676Z * [new branch] gh/oulgen/25/orig -> origin/gh/oulgen/25/orig 2025-12-04T09:16:47.6545164Z * [new branch] gh/oulgen/26/base -> origin/gh/oulgen/26/base 2025-12-04T09:16:47.6546969Z * [new branch] gh/oulgen/26/head -> origin/gh/oulgen/26/head 2025-12-04T09:16:47.6548861Z * [new branch] gh/oulgen/26/orig -> origin/gh/oulgen/26/orig 2025-12-04T09:16:47.6551319Z * [new branch] gh/oulgen/4/base -> origin/gh/oulgen/4/base 2025-12-04T09:16:47.6553179Z * [new branch] gh/oulgen/4/head -> origin/gh/oulgen/4/head 2025-12-04T09:16:47.6555054Z * [new branch] gh/oulgen/4/orig -> origin/gh/oulgen/4/orig 2025-12-04T09:16:47.6558007Z * [new branch] gh/oulgen/7/base -> origin/gh/oulgen/7/base 2025-12-04T09:16:47.6559852Z * [new branch] gh/oulgen/7/head -> origin/gh/oulgen/7/head 2025-12-04T09:16:47.6561785Z * [new branch] gh/oulgen/7/orig -> origin/gh/oulgen/7/orig 2025-12-04T09:16:47.6564428Z * [new branch] gh/oulgen/8/base -> origin/gh/oulgen/8/base 2025-12-04T09:16:47.6566997Z * [new branch] gh/oulgen/8/head -> origin/gh/oulgen/8/head 2025-12-04T09:16:47.6568797Z * [new branch] gh/oulgen/8/orig -> origin/gh/oulgen/8/orig 2025-12-04T09:16:47.6571253Z * [new branch] gh/oulgen/9/base -> origin/gh/oulgen/9/base 2025-12-04T09:16:47.6573024Z * [new branch] gh/oulgen/9/head -> origin/gh/oulgen/9/head 2025-12-04T09:16:47.6574927Z * [new branch] gh/oulgen/9/orig -> origin/gh/oulgen/9/orig 2025-12-04T09:16:47.6577456Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-12-04T09:16:47.6580599Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-12-04T09:16:47.6582506Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-12-04T09:16:47.6584604Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-12-04T09:16:47.6587037Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-12-04T09:16:47.6588800Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-12-04T09:16:47.6590589Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-12-04T09:16:47.6593141Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-12-04T09:16:47.6594951Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-12-04T09:16:47.6596705Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-12-04T09:16:47.6599280Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-12-04T09:16:47.6601019Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-12-04T09:16:47.6603367Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-12-04T09:16:47.6605830Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-12-04T09:16:47.6607789Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-12-04T09:16:47.6609928Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-12-04T09:16:47.6612313Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-12-04T09:16:47.6614255Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-12-04T09:16:47.6616095Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-12-04T09:16:47.6618441Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-12-04T09:16:47.6620313Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-12-04T09:16:47.6622205Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-12-04T09:16:47.6624649Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-12-04T09:16:47.6626637Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-12-04T09:16:47.6628321Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-12-04T09:16:47.6630838Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-12-04T09:16:47.6632695Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-12-04T09:16:47.6634516Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-12-04T09:16:47.6640366Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-12-04T09:16:47.6642077Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-12-04T09:16:47.6643935Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-12-04T09:16:47.6646473Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-12-04T09:16:47.6648313Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-12-04T09:16:47.6650196Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-12-04T09:16:47.6652702Z * [new branch] gh/pearu/140/base -> origin/gh/pearu/140/base 2025-12-04T09:16:47.6654830Z * [new branch] gh/pearu/140/head -> origin/gh/pearu/140/head 2025-12-04T09:16:47.6656487Z * [new branch] gh/pearu/140/orig -> origin/gh/pearu/140/orig 2025-12-04T09:16:47.6659058Z * [new branch] gh/pearu/142/base -> origin/gh/pearu/142/base 2025-12-04T09:16:47.6660862Z * [new branch] gh/pearu/142/head -> origin/gh/pearu/142/head 2025-12-04T09:16:47.6662643Z * [new branch] gh/pearu/142/orig -> origin/gh/pearu/142/orig 2025-12-04T09:16:47.6665210Z * [new branch] gh/pearu/143/base -> origin/gh/pearu/143/base 2025-12-04T09:16:47.6667025Z * [new branch] gh/pearu/143/head -> origin/gh/pearu/143/head 2025-12-04T09:16:47.6668949Z * [new branch] gh/pearu/143/orig -> origin/gh/pearu/143/orig 2025-12-04T09:16:47.6671863Z * [new branch] gh/pearu/147/base -> origin/gh/pearu/147/base 2025-12-04T09:16:47.6673788Z * [new branch] gh/pearu/147/head -> origin/gh/pearu/147/head 2025-12-04T09:16:47.6675503Z * [new branch] gh/pearu/147/orig -> origin/gh/pearu/147/orig 2025-12-04T09:16:47.6678095Z * [new branch] gh/pearu/149/base -> origin/gh/pearu/149/base 2025-12-04T09:16:47.6679899Z * [new branch] gh/pearu/149/head -> origin/gh/pearu/149/head 2025-12-04T09:16:47.6681643Z * [new branch] gh/pearu/149/orig -> origin/gh/pearu/149/orig 2025-12-04T09:16:47.6684704Z * [new branch] gh/pearu/150/base -> origin/gh/pearu/150/base 2025-12-04T09:16:47.6686491Z * [new branch] gh/pearu/150/head -> origin/gh/pearu/150/head 2025-12-04T09:16:47.6688524Z * [new branch] gh/pearu/150/orig -> origin/gh/pearu/150/orig 2025-12-04T09:16:47.6691963Z * [new branch] gh/pearu/151/base -> origin/gh/pearu/151/base 2025-12-04T09:16:47.6694372Z * [new branch] gh/pearu/151/head -> origin/gh/pearu/151/head 2025-12-04T09:16:47.6696133Z * [new branch] gh/pearu/151/orig -> origin/gh/pearu/151/orig 2025-12-04T09:16:47.6698814Z * [new branch] gh/pearu/152/base -> origin/gh/pearu/152/base 2025-12-04T09:16:47.6700686Z * [new branch] gh/pearu/152/head -> origin/gh/pearu/152/head 2025-12-04T09:16:47.6702480Z * [new branch] gh/pearu/152/orig -> origin/gh/pearu/152/orig 2025-12-04T09:16:47.6704988Z * [new branch] gh/pearu/153/base -> origin/gh/pearu/153/base 2025-12-04T09:16:47.6706913Z * [new branch] gh/pearu/153/head -> origin/gh/pearu/153/head 2025-12-04T09:16:47.6708623Z * [new branch] gh/pearu/153/orig -> origin/gh/pearu/153/orig 2025-12-04T09:16:47.6711735Z * [new branch] gh/pearu/154/base -> origin/gh/pearu/154/base 2025-12-04T09:16:47.6713678Z * [new branch] gh/pearu/154/head -> origin/gh/pearu/154/head 2025-12-04T09:16:47.6715406Z * [new branch] gh/pearu/154/orig -> origin/gh/pearu/154/orig 2025-12-04T09:16:47.6718154Z * [new branch] gh/pearu/155/base -> origin/gh/pearu/155/base 2025-12-04T09:16:47.6720133Z * [new branch] gh/pearu/155/head -> origin/gh/pearu/155/head 2025-12-04T09:16:47.6721810Z * [new branch] gh/pearu/155/orig -> origin/gh/pearu/155/orig 2025-12-04T09:16:47.6724278Z * [new branch] gh/pearu/156/base -> origin/gh/pearu/156/base 2025-12-04T09:16:47.6726087Z * [new branch] gh/pearu/156/head -> origin/gh/pearu/156/head 2025-12-04T09:16:47.6728006Z * [new branch] gh/pearu/156/orig -> origin/gh/pearu/156/orig 2025-12-04T09:16:47.6731138Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-12-04T09:16:47.6733326Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-12-04T09:16:47.6735000Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-12-04T09:16:47.6738387Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-12-04T09:16:47.6740293Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-12-04T09:16:47.6742200Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-12-04T09:16:47.6745419Z * [new branch] gh/pianpwk/21/base -> origin/gh/pianpwk/21/base 2025-12-04T09:16:47.6747309Z * [new branch] gh/pianpwk/21/head -> origin/gh/pianpwk/21/head 2025-12-04T09:16:47.6749938Z * [new branch] gh/pianpwk/28/base -> origin/gh/pianpwk/28/base 2025-12-04T09:16:47.6752249Z * [new branch] gh/pianpwk/28/head -> origin/gh/pianpwk/28/head 2025-12-04T09:16:47.6753971Z * [new branch] gh/pianpwk/28/orig -> origin/gh/pianpwk/28/orig 2025-12-04T09:16:47.6756547Z * [new branch] gh/pianpwk/29/base -> origin/gh/pianpwk/29/base 2025-12-04T09:16:47.6758760Z * [new branch] gh/pianpwk/29/head -> origin/gh/pianpwk/29/head 2025-12-04T09:16:47.6760552Z * [new branch] gh/pianpwk/29/orig -> origin/gh/pianpwk/29/orig 2025-12-04T09:16:47.6763264Z * [new branch] gh/pianpwk/30/base -> origin/gh/pianpwk/30/base 2025-12-04T09:16:47.6765104Z * [new branch] gh/pianpwk/30/head -> origin/gh/pianpwk/30/head 2025-12-04T09:16:47.6766971Z * [new branch] gh/pianpwk/30/orig -> origin/gh/pianpwk/30/orig 2025-12-04T09:16:47.6769766Z * [new branch] gh/pianpwk/31/base -> origin/gh/pianpwk/31/base 2025-12-04T09:16:47.6771571Z * [new branch] gh/pianpwk/31/head -> origin/gh/pianpwk/31/head 2025-12-04T09:16:47.6773440Z * [new branch] gh/pianpwk/31/orig -> origin/gh/pianpwk/31/orig 2025-12-04T09:16:47.6775761Z * [new branch] gh/pianpwk/32/base -> origin/gh/pianpwk/32/base 2025-12-04T09:16:47.6777534Z * [new branch] gh/pianpwk/32/head -> origin/gh/pianpwk/32/head 2025-12-04T09:16:47.6779454Z * [new branch] gh/pianpwk/32/orig -> origin/gh/pianpwk/32/orig 2025-12-04T09:16:47.6781709Z * [new branch] gh/pianpwk/33/base -> origin/gh/pianpwk/33/base 2025-12-04T09:16:47.6783573Z * [new branch] gh/pianpwk/33/head -> origin/gh/pianpwk/33/head 2025-12-04T09:16:47.6785568Z * [new branch] gh/pianpwk/33/orig -> origin/gh/pianpwk/33/orig 2025-12-04T09:16:47.6788146Z * [new branch] gh/pianpwk/34/base -> origin/gh/pianpwk/34/base 2025-12-04T09:16:47.6790292Z * [new branch] gh/pianpwk/34/head -> origin/gh/pianpwk/34/head 2025-12-04T09:16:47.6792247Z * [new branch] gh/pianpwk/34/orig -> origin/gh/pianpwk/34/orig 2025-12-04T09:16:47.6794827Z * [new branch] gh/pianpwk/35/base -> origin/gh/pianpwk/35/base 2025-12-04T09:16:47.6796800Z * [new branch] gh/pianpwk/35/head -> origin/gh/pianpwk/35/head 2025-12-04T09:16:47.6798839Z * [new branch] gh/pianpwk/35/orig -> origin/gh/pianpwk/35/orig 2025-12-04T09:16:47.6801871Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-12-04T09:16:47.6803681Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-12-04T09:16:47.6806103Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-12-04T09:16:47.6808599Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-12-04T09:16:47.6810454Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-12-04T09:16:47.6813092Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-12-04T09:16:47.6814754Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-12-04T09:16:47.6816557Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-12-04T09:16:47.6819197Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-12-04T09:16:47.6820997Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-12-04T09:16:47.6822766Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-12-04T09:16:47.6825288Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-12-04T09:16:47.6827128Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-12-04T09:16:47.6828948Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-12-04T09:16:47.6831450Z * [new branch] gh/rec/167/base -> origin/gh/rec/167/base 2025-12-04T09:16:47.6833268Z * [new branch] gh/rec/167/head -> origin/gh/rec/167/head 2025-12-04T09:16:47.6835100Z * [new branch] gh/rec/167/orig -> origin/gh/rec/167/orig 2025-12-04T09:16:47.6837734Z * [new branch] gh/rec/168/base -> origin/gh/rec/168/base 2025-12-04T09:16:47.6839604Z * [new branch] gh/rec/168/head -> origin/gh/rec/168/head 2025-12-04T09:16:47.6841356Z * [new branch] gh/rec/168/orig -> origin/gh/rec/168/orig 2025-12-04T09:16:47.6843996Z * [new branch] gh/rec/169/base -> origin/gh/rec/169/base 2025-12-04T09:16:47.6845794Z * [new branch] gh/rec/169/head -> origin/gh/rec/169/head 2025-12-04T09:16:47.6847678Z * [new branch] gh/rec/169/orig -> origin/gh/rec/169/orig 2025-12-04T09:16:47.6850270Z * [new branch] gh/rec/170/base -> origin/gh/rec/170/base 2025-12-04T09:16:47.6852080Z * [new branch] gh/rec/170/head -> origin/gh/rec/170/head 2025-12-04T09:16:47.6853855Z * [new branch] gh/rec/170/orig -> origin/gh/rec/170/orig 2025-12-04T09:16:47.6856410Z * [new branch] gh/rec/171/base -> origin/gh/rec/171/base 2025-12-04T09:16:47.6858808Z * [new branch] gh/rec/171/head -> origin/gh/rec/171/head 2025-12-04T09:16:47.6860677Z * [new branch] gh/rec/171/orig -> origin/gh/rec/171/orig 2025-12-04T09:16:47.6863125Z * [new branch] gh/rec/172/base -> origin/gh/rec/172/base 2025-12-04T09:16:47.6864966Z * [new branch] gh/rec/172/head -> origin/gh/rec/172/head 2025-12-04T09:16:47.6866693Z * [new branch] gh/rec/172/orig -> origin/gh/rec/172/orig 2025-12-04T09:16:47.6869341Z * [new branch] gh/rec/173/base -> origin/gh/rec/173/base 2025-12-04T09:16:47.6871123Z * [new branch] gh/rec/173/head -> origin/gh/rec/173/head 2025-12-04T09:16:47.6872919Z * [new branch] gh/rec/173/orig -> origin/gh/rec/173/orig 2025-12-04T09:16:47.6875460Z * [new branch] gh/rec/174/base -> origin/gh/rec/174/base 2025-12-04T09:16:47.6877267Z * [new branch] gh/rec/174/head -> origin/gh/rec/174/head 2025-12-04T09:16:47.6879138Z * [new branch] gh/rec/174/orig -> origin/gh/rec/174/orig 2025-12-04T09:16:47.6882071Z * [new branch] gh/rec/175/base -> origin/gh/rec/175/base 2025-12-04T09:16:47.6883957Z * [new branch] gh/rec/175/head -> origin/gh/rec/175/head 2025-12-04T09:16:47.6885713Z * [new branch] gh/rec/175/orig -> origin/gh/rec/175/orig 2025-12-04T09:16:47.6888486Z * [new branch] gh/rec/176/base -> origin/gh/rec/176/base 2025-12-04T09:16:47.6890184Z * [new branch] gh/rec/176/head -> origin/gh/rec/176/head 2025-12-04T09:16:47.6892008Z * [new branch] gh/rec/176/orig -> origin/gh/rec/176/orig 2025-12-04T09:16:47.6894518Z * [new branch] gh/rec/177/base -> origin/gh/rec/177/base 2025-12-04T09:16:47.6896399Z * [new branch] gh/rec/177/head -> origin/gh/rec/177/head 2025-12-04T09:16:47.6898228Z * [new branch] gh/rec/177/orig -> origin/gh/rec/177/orig 2025-12-04T09:16:47.6901422Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-12-04T09:16:47.6903192Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-12-04T09:16:47.6905036Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-12-04T09:16:47.6907707Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-12-04T09:16:47.6909895Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-12-04T09:16:47.6911679Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-12-04T09:16:47.6914234Z * [new branch] gh/robert-hardwick/5/base -> origin/gh/robert-hardwick/5/base 2025-12-04T09:16:47.6916634Z * [new branch] gh/robert-hardwick/5/head -> origin/gh/robert-hardwick/5/head 2025-12-04T09:16:47.6918599Z * [new branch] gh/robert-hardwick/5/orig -> origin/gh/robert-hardwick/5/orig 2025-12-04T09:16:47.6921122Z * [new branch] gh/robert-hardwick/6/base -> origin/gh/robert-hardwick/6/base 2025-12-04T09:16:47.6922916Z * [new branch] gh/robert-hardwick/6/head -> origin/gh/robert-hardwick/6/head 2025-12-04T09:16:47.6924734Z * [new branch] gh/robert-hardwick/6/orig -> origin/gh/robert-hardwick/6/orig 2025-12-04T09:16:47.6927241Z * [new branch] gh/robert-hardwick/7/base -> origin/gh/robert-hardwick/7/base 2025-12-04T09:16:47.6929288Z * [new branch] gh/robert-hardwick/7/head -> origin/gh/robert-hardwick/7/head 2025-12-04T09:16:47.6931062Z * [new branch] gh/robert-hardwick/7/orig -> origin/gh/robert-hardwick/7/orig 2025-12-04T09:16:47.6933958Z * [new branch] gh/robert-hardwick/8/base -> origin/gh/robert-hardwick/8/base 2025-12-04T09:16:47.6936030Z * [new branch] gh/robert-hardwick/8/head -> origin/gh/robert-hardwick/8/head 2025-12-04T09:16:47.6938166Z * [new branch] gh/robert-hardwick/8/orig -> origin/gh/robert-hardwick/8/orig 2025-12-04T09:16:47.6940619Z * [new branch] gh/robert-hardwick/9/base -> origin/gh/robert-hardwick/9/base 2025-12-04T09:16:47.6942455Z * [new branch] gh/robert-hardwick/9/head -> origin/gh/robert-hardwick/9/head 2025-12-04T09:16:47.6944394Z * [new branch] gh/robert-hardwick/9/orig -> origin/gh/robert-hardwick/9/orig 2025-12-04T09:16:47.6947416Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-12-04T09:16:47.6949380Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-12-04T09:16:47.6951855Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-12-04T09:16:47.6953618Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-12-04T09:16:47.6956151Z * [new branch] gh/rtimpe/22/base -> origin/gh/rtimpe/22/base 2025-12-04T09:16:47.6957984Z * [new branch] gh/rtimpe/22/head -> origin/gh/rtimpe/22/head 2025-12-04T09:16:47.6959765Z * [new branch] gh/rtimpe/22/orig -> origin/gh/rtimpe/22/orig 2025-12-04T09:16:47.6962115Z * [new branch] gh/rtimpe/23/base -> origin/gh/rtimpe/23/base 2025-12-04T09:16:47.6964118Z * [new branch] gh/rtimpe/23/head -> origin/gh/rtimpe/23/head 2025-12-04T09:16:47.6965806Z * [new branch] gh/rtimpe/23/orig -> origin/gh/rtimpe/23/orig 2025-12-04T09:16:47.6968469Z * [new branch] gh/rtimpe/24/base -> origin/gh/rtimpe/24/base 2025-12-04T09:16:47.6970330Z * [new branch] gh/rtimpe/24/head -> origin/gh/rtimpe/24/head 2025-12-04T09:16:47.6972097Z * [new branch] gh/rtimpe/24/orig -> origin/gh/rtimpe/24/orig 2025-12-04T09:16:47.6974583Z * [new branch] gh/rtimpe/25/base -> origin/gh/rtimpe/25/base 2025-12-04T09:16:47.6976477Z * [new branch] gh/rtimpe/25/head -> origin/gh/rtimpe/25/head 2025-12-04T09:16:47.6978314Z * [new branch] gh/rtimpe/25/orig -> origin/gh/rtimpe/25/orig 2025-12-04T09:16:47.6980732Z * [new branch] gh/rtimpe/26/base -> origin/gh/rtimpe/26/base 2025-12-04T09:16:47.6982589Z * [new branch] gh/rtimpe/26/head -> origin/gh/rtimpe/26/head 2025-12-04T09:16:47.6984487Z * [new branch] gh/rtimpe/26/orig -> origin/gh/rtimpe/26/orig 2025-12-04T09:16:47.6986814Z * [new branch] gh/rtimpe/27/base -> origin/gh/rtimpe/27/base 2025-12-04T09:16:47.6988633Z * [new branch] gh/rtimpe/27/head -> origin/gh/rtimpe/27/head 2025-12-04T09:16:47.6990515Z * [new branch] gh/rtimpe/27/orig -> origin/gh/rtimpe/27/orig 2025-12-04T09:16:47.6993114Z * [new branch] gh/rtimpe/28/base -> origin/gh/rtimpe/28/base 2025-12-04T09:16:47.6994908Z * [new branch] gh/rtimpe/28/head -> origin/gh/rtimpe/28/head 2025-12-04T09:16:47.6996769Z * [new branch] gh/rtimpe/28/orig -> origin/gh/rtimpe/28/orig 2025-12-04T09:16:47.6999468Z * [new branch] gh/rtimpe/29/base -> origin/gh/rtimpe/29/base 2025-12-04T09:16:47.7001089Z * [new branch] gh/rtimpe/29/head -> origin/gh/rtimpe/29/head 2025-12-04T09:16:47.7002865Z * [new branch] gh/rtimpe/29/orig -> origin/gh/rtimpe/29/orig 2025-12-04T09:16:47.7005245Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-12-04T09:16:47.7007101Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-12-04T09:16:47.7009861Z * [new branch] gh/rtimpe/30/base -> origin/gh/rtimpe/30/base 2025-12-04T09:16:47.7011611Z * [new branch] gh/rtimpe/30/head -> origin/gh/rtimpe/30/head 2025-12-04T09:16:47.7013387Z * [new branch] gh/rtimpe/30/orig -> origin/gh/rtimpe/30/orig 2025-12-04T09:16:47.7015942Z * [new branch] gh/rtimpe/31/base -> origin/gh/rtimpe/31/base 2025-12-04T09:16:47.7017816Z * [new branch] gh/rtimpe/31/head -> origin/gh/rtimpe/31/head 2025-12-04T09:16:47.7019864Z * [new branch] gh/rtimpe/31/orig -> origin/gh/rtimpe/31/orig 2025-12-04T09:16:47.7022159Z * [new branch] gh/rtimpe/32/base -> origin/gh/rtimpe/32/base 2025-12-04T09:16:47.7023975Z * [new branch] gh/rtimpe/32/head -> origin/gh/rtimpe/32/head 2025-12-04T09:16:47.7025719Z * [new branch] gh/rtimpe/32/orig -> origin/gh/rtimpe/32/orig 2025-12-04T09:16:47.7028196Z * [new branch] gh/rtimpe/33/base -> origin/gh/rtimpe/33/base 2025-12-04T09:16:47.7030005Z * [new branch] gh/rtimpe/33/head -> origin/gh/rtimpe/33/head 2025-12-04T09:16:47.7031823Z * [new branch] gh/rtimpe/33/orig -> origin/gh/rtimpe/33/orig 2025-12-04T09:16:47.7034251Z * [new branch] gh/rtimpe/34/base -> origin/gh/rtimpe/34/base 2025-12-04T09:16:47.7036268Z * [new branch] gh/rtimpe/34/head -> origin/gh/rtimpe/34/head 2025-12-04T09:16:47.7039574Z * [new branch] gh/rtimpe/34/orig -> origin/gh/rtimpe/34/orig 2025-12-04T09:16:47.7042029Z * [new branch] gh/rtimpe/35/base -> origin/gh/rtimpe/35/base 2025-12-04T09:16:47.7043919Z * [new branch] gh/rtimpe/35/head -> origin/gh/rtimpe/35/head 2025-12-04T09:16:47.7045785Z * [new branch] gh/rtimpe/35/orig -> origin/gh/rtimpe/35/orig 2025-12-04T09:16:47.7048315Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-12-04T09:16:47.7050160Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-12-04T09:16:47.7053203Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-12-04T09:16:47.7055256Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-12-04T09:16:47.7056913Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-12-04T09:16:47.7059436Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-12-04T09:16:47.7061273Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-12-04T09:16:47.7063115Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-12-04T09:16:47.7065791Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-12-04T09:16:47.7067615Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-12-04T09:16:47.7069387Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-12-04T09:16:47.7071867Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-12-04T09:16:47.7073664Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-12-04T09:16:47.7075474Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-12-04T09:16:47.7078019Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-12-04T09:16:47.7080394Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-12-04T09:16:47.7082165Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-12-04T09:16:47.7084765Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-12-04T09:16:47.7086910Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-12-04T09:16:47.7088648Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-12-04T09:16:47.7091280Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-12-04T09:16:47.7093290Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-12-04T09:16:47.7094924Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-12-04T09:16:47.7097993Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-12-04T09:16:47.7099882Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-12-04T09:16:47.7101749Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-12-04T09:16:47.7104204Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-12-04T09:16:47.7106195Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-12-04T09:16:47.7107989Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-12-04T09:16:47.7110470Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-12-04T09:16:47.7112289Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-12-04T09:16:47.7114218Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-12-04T09:16:47.7116646Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-12-04T09:16:47.7118405Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-12-04T09:16:47.7120225Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-12-04T09:16:47.7122560Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-12-04T09:16:47.7124402Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-12-04T09:16:47.7126250Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-12-04T09:16:47.7128918Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-12-04T09:16:47.7130762Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-12-04T09:16:47.7132549Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-12-04T09:16:47.7135045Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-12-04T09:16:47.7137385Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-12-04T09:16:47.7139209Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-12-04T09:16:47.7141744Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-12-04T09:16:47.7143525Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-12-04T09:16:47.7145326Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-12-04T09:16:47.7147885Z * [new branch] gh/seemethere/72/base -> origin/gh/seemethere/72/base 2025-12-04T09:16:47.7149714Z * [new branch] gh/seemethere/72/head -> origin/gh/seemethere/72/head 2025-12-04T09:16:47.7151516Z * [new branch] gh/seemethere/72/orig -> origin/gh/seemethere/72/orig 2025-12-04T09:16:47.7154006Z * [new branch] gh/seemethere/73/base -> origin/gh/seemethere/73/base 2025-12-04T09:16:47.7155792Z * [new branch] gh/seemethere/73/head -> origin/gh/seemethere/73/head 2025-12-04T09:16:47.7157617Z * [new branch] gh/seemethere/73/orig -> origin/gh/seemethere/73/orig 2025-12-04T09:16:47.7160166Z * [new branch] gh/seemethere/74/base -> origin/gh/seemethere/74/base 2025-12-04T09:16:47.7161902Z * [new branch] gh/seemethere/74/head -> origin/gh/seemethere/74/head 2025-12-04T09:16:47.7163786Z * [new branch] gh/seemethere/74/orig -> origin/gh/seemethere/74/orig 2025-12-04T09:16:47.7166397Z * [new branch] gh/seemethere/75/base -> origin/gh/seemethere/75/base 2025-12-04T09:16:47.7168421Z * [new branch] gh/seemethere/75/head -> origin/gh/seemethere/75/head 2025-12-04T09:16:47.7170356Z * [new branch] gh/seemethere/75/orig -> origin/gh/seemethere/75/orig 2025-12-04T09:16:47.7174030Z * [new branch] gh/seemethere/76/base -> origin/gh/seemethere/76/base 2025-12-04T09:16:47.7175922Z * [new branch] gh/seemethere/76/head -> origin/gh/seemethere/76/head 2025-12-04T09:16:47.7177758Z * [new branch] gh/seemethere/76/orig -> origin/gh/seemethere/76/orig 2025-12-04T09:16:47.7181098Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-12-04T09:16:47.7183092Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-12-04T09:16:47.7185002Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-12-04T09:16:47.7187784Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-12-04T09:16:47.7189747Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-12-04T09:16:47.7191685Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-12-04T09:16:47.7194263Z * [new branch] gh/shunting314/249/base -> origin/gh/shunting314/249/base 2025-12-04T09:16:47.7196152Z * [new branch] gh/shunting314/249/head -> origin/gh/shunting314/249/head 2025-12-04T09:16:47.7198077Z * [new branch] gh/shunting314/249/orig -> origin/gh/shunting314/249/orig 2025-12-04T09:16:47.7200690Z * [new branch] gh/shunting314/253/base -> origin/gh/shunting314/253/base 2025-12-04T09:16:47.7202348Z * [new branch] gh/shunting314/253/head -> origin/gh/shunting314/253/head 2025-12-04T09:16:47.7204114Z * [new branch] gh/shunting314/253/orig -> origin/gh/shunting314/253/orig 2025-12-04T09:16:47.7206693Z * [new branch] gh/shunting314/256/base -> origin/gh/shunting314/256/base 2025-12-04T09:16:47.7208673Z * [new branch] gh/shunting314/256/head -> origin/gh/shunting314/256/head 2025-12-04T09:16:47.7210462Z * [new branch] gh/shunting314/256/orig -> origin/gh/shunting314/256/orig 2025-12-04T09:16:47.7213227Z * [new branch] gh/shunting314/257/base -> origin/gh/shunting314/257/base 2025-12-04T09:16:47.7215103Z * [new branch] gh/shunting314/257/head -> origin/gh/shunting314/257/head 2025-12-04T09:16:47.7217045Z * [new branch] gh/shunting314/257/orig -> origin/gh/shunting314/257/orig 2025-12-04T09:16:47.7219687Z * [new branch] gh/shunting314/258/base -> origin/gh/shunting314/258/base 2025-12-04T09:16:47.7221482Z * [new branch] gh/shunting314/258/head -> origin/gh/shunting314/258/head 2025-12-04T09:16:47.7223331Z * [new branch] gh/shunting314/258/orig -> origin/gh/shunting314/258/orig 2025-12-04T09:16:47.7225716Z * [new branch] gh/shunting314/259/base -> origin/gh/shunting314/259/base 2025-12-04T09:16:47.7227657Z * [new branch] gh/shunting314/259/head -> origin/gh/shunting314/259/head 2025-12-04T09:16:47.7229467Z * [new branch] gh/shunting314/259/orig -> origin/gh/shunting314/259/orig 2025-12-04T09:16:47.7232057Z * [new branch] gh/shunting314/260/base -> origin/gh/shunting314/260/base 2025-12-04T09:16:47.7234005Z * [new branch] gh/shunting314/260/head -> origin/gh/shunting314/260/head 2025-12-04T09:16:47.7235971Z * [new branch] gh/shunting314/260/orig -> origin/gh/shunting314/260/orig 2025-12-04T09:16:47.7239202Z * [new branch] gh/shunting314/261/base -> origin/gh/shunting314/261/base 2025-12-04T09:16:47.7241095Z * [new branch] gh/shunting314/261/head -> origin/gh/shunting314/261/head 2025-12-04T09:16:47.7243057Z * [new branch] gh/shunting314/261/orig -> origin/gh/shunting314/261/orig 2025-12-04T09:16:47.7245624Z * [new branch] gh/shunting314/262/base -> origin/gh/shunting314/262/base 2025-12-04T09:16:47.7247612Z * [new branch] gh/shunting314/262/head -> origin/gh/shunting314/262/head 2025-12-04T09:16:47.7249500Z * [new branch] gh/shunting314/262/orig -> origin/gh/shunting314/262/orig 2025-12-04T09:16:47.7252288Z * [new branch] gh/shunting314/263/base -> origin/gh/shunting314/263/base 2025-12-04T09:16:47.7253978Z * [new branch] gh/shunting314/263/head -> origin/gh/shunting314/263/head 2025-12-04T09:16:47.7255801Z * [new branch] gh/shunting314/263/orig -> origin/gh/shunting314/263/orig 2025-12-04T09:16:47.7258332Z * [new branch] gh/shunting314/264/base -> origin/gh/shunting314/264/base 2025-12-04T09:16:47.7260601Z * [new branch] gh/shunting314/264/head -> origin/gh/shunting314/264/head 2025-12-04T09:16:47.7262189Z * [new branch] gh/shunting314/264/orig -> origin/gh/shunting314/264/orig 2025-12-04T09:16:47.7264861Z * [new branch] gh/shunting314/265/base -> origin/gh/shunting314/265/base 2025-12-04T09:16:47.7266490Z * [new branch] gh/shunting314/265/head -> origin/gh/shunting314/265/head 2025-12-04T09:16:47.7268608Z * [new branch] gh/shunting314/265/orig -> origin/gh/shunting314/265/orig 2025-12-04T09:16:47.7271172Z * [new branch] gh/shunting314/266/base -> origin/gh/shunting314/266/base 2025-12-04T09:16:47.7273382Z * [new branch] gh/shunting314/266/head -> origin/gh/shunting314/266/head 2025-12-04T09:16:47.7275046Z * [new branch] gh/shunting314/266/orig -> origin/gh/shunting314/266/orig 2025-12-04T09:16:47.7278343Z * [new branch] gh/shunting314/267/base -> origin/gh/shunting314/267/base 2025-12-04T09:16:47.7280274Z * [new branch] gh/shunting314/267/head -> origin/gh/shunting314/267/head 2025-12-04T09:16:47.7282224Z * [new branch] gh/shunting314/267/orig -> origin/gh/shunting314/267/orig 2025-12-04T09:16:47.7285258Z * [new branch] gh/shunting314/268/base -> origin/gh/shunting314/268/base 2025-12-04T09:16:47.7287147Z * [new branch] gh/shunting314/268/head -> origin/gh/shunting314/268/head 2025-12-04T09:16:47.7289264Z * [new branch] gh/shunting314/268/orig -> origin/gh/shunting314/268/orig 2025-12-04T09:16:47.7291927Z * [new branch] gh/shunting314/269/base -> origin/gh/shunting314/269/base 2025-12-04T09:16:47.7293757Z * [new branch] gh/shunting314/269/head -> origin/gh/shunting314/269/head 2025-12-04T09:16:47.7295672Z * [new branch] gh/shunting314/269/orig -> origin/gh/shunting314/269/orig 2025-12-04T09:16:47.7298708Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-12-04T09:16:47.7300500Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-12-04T09:16:47.7302857Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-12-04T09:16:47.7304618Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-12-04T09:16:47.7306920Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-12-04T09:16:47.7308766Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-12-04T09:16:47.7311883Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-12-04T09:16:47.7313633Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-12-04T09:16:47.7316690Z * [new branch] gh/slayton58/39/base -> origin/gh/slayton58/39/base 2025-12-04T09:16:47.7318691Z * [new branch] gh/slayton58/39/head -> origin/gh/slayton58/39/head 2025-12-04T09:16:47.7320440Z * [new branch] gh/slayton58/39/orig -> origin/gh/slayton58/39/orig 2025-12-04T09:16:47.7322919Z * [new branch] gh/slayton58/42/base -> origin/gh/slayton58/42/base 2025-12-04T09:16:47.7324757Z * [new branch] gh/slayton58/42/head -> origin/gh/slayton58/42/head 2025-12-04T09:16:47.7326669Z * [new branch] gh/slayton58/42/orig -> origin/gh/slayton58/42/orig 2025-12-04T09:16:47.7329378Z * [new branch] gh/slayton58/43/base -> origin/gh/slayton58/43/base 2025-12-04T09:16:47.7331228Z * [new branch] gh/slayton58/43/head -> origin/gh/slayton58/43/head 2025-12-04T09:16:47.7333063Z * [new branch] gh/slayton58/43/orig -> origin/gh/slayton58/43/orig 2025-12-04T09:16:47.7336286Z * [new branch] gh/slayton58/44/base -> origin/gh/slayton58/44/base 2025-12-04T09:16:47.7338594Z * [new branch] gh/slayton58/44/head -> origin/gh/slayton58/44/head 2025-12-04T09:16:47.7340410Z * [new branch] gh/slayton58/44/orig -> origin/gh/slayton58/44/orig 2025-12-04T09:16:47.7342891Z * [new branch] gh/slayton58/45/base -> origin/gh/slayton58/45/base 2025-12-04T09:16:47.7344781Z * [new branch] gh/slayton58/45/head -> origin/gh/slayton58/45/head 2025-12-04T09:16:47.7346817Z * [new branch] gh/slayton58/45/orig -> origin/gh/slayton58/45/orig 2025-12-04T09:16:47.7349205Z * [new branch] gh/slayton58/46/base -> origin/gh/slayton58/46/base 2025-12-04T09:16:47.7351117Z * [new branch] gh/slayton58/46/head -> origin/gh/slayton58/46/head 2025-12-04T09:16:47.7352888Z * [new branch] gh/slayton58/46/orig -> origin/gh/slayton58/46/orig 2025-12-04T09:16:47.7355492Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-12-04T09:16:47.7357369Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-12-04T09:16:47.7359686Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-12-04T09:16:47.7361407Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-12-04T09:16:47.7364704Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-12-04T09:16:47.7366548Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-12-04T09:16:47.7368742Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-12-04T09:16:47.7371412Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-12-04T09:16:47.7373254Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-12-04T09:16:47.7375018Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-12-04T09:16:47.7377834Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-12-04T09:16:47.7379639Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-12-04T09:16:47.7381528Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-12-04T09:16:47.7384131Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-12-04T09:16:47.7386079Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-12-04T09:16:47.7387965Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-12-04T09:16:47.7390519Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-12-04T09:16:47.7392460Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-12-04T09:16:47.7394725Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-12-04T09:16:47.7397257Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-12-04T09:16:47.7399258Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-12-04T09:16:47.7401107Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-12-04T09:16:47.7403699Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-12-04T09:16:47.7405883Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-12-04T09:16:47.7407791Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-12-04T09:16:47.7410188Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-12-04T09:16:47.7412009Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-12-04T09:16:47.7414102Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-12-04T09:16:47.7417137Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-12-04T09:16:47.7419083Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-12-04T09:16:47.7420990Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-12-04T09:16:47.7423550Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-12-04T09:16:47.7425916Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-12-04T09:16:47.7427704Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-12-04T09:16:47.7430142Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-12-04T09:16:47.7432324Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-12-04T09:16:47.7433770Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-12-04T09:16:47.7436307Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-12-04T09:16:47.7440513Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-12-04T09:16:47.7442343Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-12-04T09:16:47.7445033Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-12-04T09:16:47.7446984Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-12-04T09:16:47.7449364Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-12-04T09:16:47.7451637Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-12-04T09:16:47.7453377Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-12-04T09:16:47.7455146Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-12-04T09:16:47.7457752Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-12-04T09:16:47.7459670Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-12-04T09:16:47.7461569Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-12-04T09:16:47.7463880Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-12-04T09:16:47.7465846Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-12-04T09:16:47.7467764Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-12-04T09:16:47.7470983Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-12-04T09:16:47.7472942Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-12-04T09:16:47.7474793Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-12-04T09:16:47.7477942Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-12-04T09:16:47.7479793Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-12-04T09:16:47.7481866Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-12-04T09:16:47.7484244Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-12-04T09:16:47.7486080Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-12-04T09:16:47.7488044Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-12-04T09:16:47.7490679Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-12-04T09:16:47.7492596Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-12-04T09:16:47.7494261Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-12-04T09:16:47.7496792Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-12-04T09:16:47.7498698Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-12-04T09:16:47.7500561Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-12-04T09:16:47.7503036Z * [new branch] gh/soulitzer/385/base -> origin/gh/soulitzer/385/base 2025-12-04T09:16:47.7504880Z * [new branch] gh/soulitzer/385/head -> origin/gh/soulitzer/385/head 2025-12-04T09:16:47.7506682Z * [new branch] gh/soulitzer/385/orig -> origin/gh/soulitzer/385/orig 2025-12-04T09:16:47.7509425Z * [new branch] gh/soulitzer/386/base -> origin/gh/soulitzer/386/base 2025-12-04T09:16:47.7511237Z * [new branch] gh/soulitzer/386/head -> origin/gh/soulitzer/386/head 2025-12-04T09:16:47.7513037Z * [new branch] gh/soulitzer/386/orig -> origin/gh/soulitzer/386/orig 2025-12-04T09:16:47.7515610Z * [new branch] gh/soulitzer/387/base -> origin/gh/soulitzer/387/base 2025-12-04T09:16:47.7517406Z * [new branch] gh/soulitzer/387/head -> origin/gh/soulitzer/387/head 2025-12-04T09:16:47.7519281Z * [new branch] gh/soulitzer/387/orig -> origin/gh/soulitzer/387/orig 2025-12-04T09:16:47.7521819Z * [new branch] gh/soulitzer/388/base -> origin/gh/soulitzer/388/base 2025-12-04T09:16:47.7523720Z * [new branch] gh/soulitzer/388/head -> origin/gh/soulitzer/388/head 2025-12-04T09:16:47.7525834Z * [new branch] gh/soulitzer/388/orig -> origin/gh/soulitzer/388/orig 2025-12-04T09:16:47.7528355Z * [new branch] gh/soulitzer/389/base -> origin/gh/soulitzer/389/base 2025-12-04T09:16:47.7530150Z * [new branch] gh/soulitzer/389/head -> origin/gh/soulitzer/389/head 2025-12-04T09:16:47.7531945Z * [new branch] gh/soulitzer/389/orig -> origin/gh/soulitzer/389/orig 2025-12-04T09:16:47.7535032Z * [new branch] gh/soulitzer/390/base -> origin/gh/soulitzer/390/base 2025-12-04T09:16:47.7537219Z * [new branch] gh/soulitzer/390/head -> origin/gh/soulitzer/390/head 2025-12-04T09:16:47.7539092Z * [new branch] gh/soulitzer/390/orig -> origin/gh/soulitzer/390/orig 2025-12-04T09:16:47.7541697Z * [new branch] gh/soulitzer/391/base -> origin/gh/soulitzer/391/base 2025-12-04T09:16:47.7543502Z * [new branch] gh/soulitzer/391/head -> origin/gh/soulitzer/391/head 2025-12-04T09:16:47.7545276Z * [new branch] gh/soulitzer/391/orig -> origin/gh/soulitzer/391/orig 2025-12-04T09:16:47.7548107Z * [new branch] gh/soulitzer/392/base -> origin/gh/soulitzer/392/base 2025-12-04T09:16:47.7549762Z * [new branch] gh/soulitzer/392/head -> origin/gh/soulitzer/392/head 2025-12-04T09:16:47.7551583Z * [new branch] gh/soulitzer/392/orig -> origin/gh/soulitzer/392/orig 2025-12-04T09:16:47.7554662Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-12-04T09:16:47.7557438Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-12-04T09:16:47.7559190Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-12-04T09:16:47.7561066Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-12-04T09:16:47.7563537Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-12-04T09:16:47.7565682Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-12-04T09:16:47.7567195Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-12-04T09:16:47.7569955Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-12-04T09:16:47.7571693Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-12-04T09:16:47.7573500Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-12-04T09:16:47.7576089Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-12-04T09:16:47.7577905Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-12-04T09:16:47.7579677Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-12-04T09:16:47.7582214Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-12-04T09:16:47.7584075Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-12-04T09:16:47.7585958Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-12-04T09:16:47.7588492Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-12-04T09:16:47.7590936Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-12-04T09:16:47.7592802Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-12-04T09:16:47.7595231Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-12-04T09:16:47.7597027Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-12-04T09:16:47.7599102Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-12-04T09:16:47.7601615Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-12-04T09:16:47.7603540Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-12-04T09:16:47.7605430Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-12-04T09:16:47.7607982Z * [new branch] gh/swolchok/856/base -> origin/gh/swolchok/856/base 2025-12-04T09:16:47.7609884Z * [new branch] gh/swolchok/856/head -> origin/gh/swolchok/856/head 2025-12-04T09:16:47.7611666Z * [new branch] gh/swolchok/856/orig -> origin/gh/swolchok/856/orig 2025-12-04T09:16:47.7614706Z * [new branch] gh/swolchok/860/base -> origin/gh/swolchok/860/base 2025-12-04T09:16:47.7616537Z * [new branch] gh/swolchok/860/head -> origin/gh/swolchok/860/head 2025-12-04T09:16:47.7618301Z * [new branch] gh/swolchok/860/orig -> origin/gh/swolchok/860/orig 2025-12-04T09:16:47.7621057Z * [new branch] gh/swolchok/861/base -> origin/gh/swolchok/861/base 2025-12-04T09:16:47.7622945Z * [new branch] gh/swolchok/861/head -> origin/gh/swolchok/861/head 2025-12-04T09:16:47.7624966Z * [new branch] gh/swolchok/861/orig -> origin/gh/swolchok/861/orig 2025-12-04T09:16:47.7627635Z * [new branch] gh/swolchok/862/base -> origin/gh/swolchok/862/base 2025-12-04T09:16:47.7629400Z * [new branch] gh/swolchok/862/head -> origin/gh/swolchok/862/head 2025-12-04T09:16:47.7631106Z * [new branch] gh/swolchok/862/orig -> origin/gh/swolchok/862/orig 2025-12-04T09:16:47.7633788Z * [new branch] gh/swolchok/863/base -> origin/gh/swolchok/863/base 2025-12-04T09:16:47.7635976Z * [new branch] gh/swolchok/863/head -> origin/gh/swolchok/863/head 2025-12-04T09:16:47.7638075Z * [new branch] gh/swolchok/863/orig -> origin/gh/swolchok/863/orig 2025-12-04T09:16:47.7640739Z * [new branch] gh/swolchok/864/base -> origin/gh/swolchok/864/base 2025-12-04T09:16:47.7642428Z * [new branch] gh/swolchok/864/head -> origin/gh/swolchok/864/head 2025-12-04T09:16:47.7644359Z * [new branch] gh/swolchok/864/orig -> origin/gh/swolchok/864/orig 2025-12-04T09:16:47.7647175Z * [new branch] gh/swolchok/865/base -> origin/gh/swolchok/865/base 2025-12-04T09:16:47.7649090Z * [new branch] gh/swolchok/865/head -> origin/gh/swolchok/865/head 2025-12-04T09:16:47.7651027Z * [new branch] gh/swolchok/865/orig -> origin/gh/swolchok/865/orig 2025-12-04T09:16:47.7654115Z * [new branch] gh/swolchok/866/base -> origin/gh/swolchok/866/base 2025-12-04T09:16:47.7655983Z * [new branch] gh/swolchok/866/head -> origin/gh/swolchok/866/head 2025-12-04T09:16:47.7657849Z * [new branch] gh/swolchok/866/orig -> origin/gh/swolchok/866/orig 2025-12-04T09:16:47.7660366Z * [new branch] gh/swolchok/867/base -> origin/gh/swolchok/867/base 2025-12-04T09:16:47.7662262Z * [new branch] gh/swolchok/867/head -> origin/gh/swolchok/867/head 2025-12-04T09:16:47.7664040Z * [new branch] gh/swolchok/867/orig -> origin/gh/swolchok/867/orig 2025-12-04T09:16:47.7666618Z * [new branch] gh/swolchok/868/base -> origin/gh/swolchok/868/base 2025-12-04T09:16:47.7668445Z * [new branch] gh/swolchok/868/head -> origin/gh/swolchok/868/head 2025-12-04T09:16:47.7670314Z * [new branch] gh/swolchok/868/orig -> origin/gh/swolchok/868/orig 2025-12-04T09:16:47.7672886Z * [new branch] gh/swolchok/869/base -> origin/gh/swolchok/869/base 2025-12-04T09:16:47.7674695Z * [new branch] gh/swolchok/869/head -> origin/gh/swolchok/869/head 2025-12-04T09:16:47.7676807Z * [new branch] gh/swolchok/869/orig -> origin/gh/swolchok/869/orig 2025-12-04T09:16:47.7679577Z * [new branch] gh/swolchok/870/base -> origin/gh/swolchok/870/base 2025-12-04T09:16:47.7681254Z * [new branch] gh/swolchok/870/head -> origin/gh/swolchok/870/head 2025-12-04T09:16:47.7683136Z * [new branch] gh/swolchok/870/orig -> origin/gh/swolchok/870/orig 2025-12-04T09:16:47.7685680Z * [new branch] gh/swolchok/871/base -> origin/gh/swolchok/871/base 2025-12-04T09:16:47.7687678Z * [new branch] gh/swolchok/871/head -> origin/gh/swolchok/871/head 2025-12-04T09:16:47.7689649Z * [new branch] gh/swolchok/871/orig -> origin/gh/swolchok/871/orig 2025-12-04T09:16:47.7692762Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-12-04T09:16:47.7694569Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-12-04T09:16:47.7696452Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-12-04T09:16:47.7699642Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-12-04T09:16:47.7701521Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-12-04T09:16:47.7703378Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-12-04T09:16:47.7705919Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-12-04T09:16:47.7707801Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-12-04T09:16:47.7710734Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-12-04T09:16:47.7712522Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-12-04T09:16:47.7714434Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-12-04T09:16:47.7718000Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-12-04T09:16:47.7719900Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-12-04T09:16:47.7721694Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-12-04T09:16:47.7724355Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-12-04T09:16:47.7726209Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-12-04T09:16:47.7728143Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-12-04T09:16:47.7730861Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-12-04T09:16:47.7732572Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-12-04T09:16:47.7734452Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-12-04T09:16:47.7737555Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-12-04T09:16:47.7739341Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-12-04T09:16:47.7741193Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-12-04T09:16:47.7743951Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-12-04T09:16:47.7745761Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-12-04T09:16:47.7747610Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-12-04T09:16:47.7750523Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-12-04T09:16:47.7752372Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-12-04T09:16:47.7754101Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-12-04T09:16:47.7756715Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-12-04T09:16:47.7758632Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-12-04T09:16:47.7760479Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-12-04T09:16:47.7763024Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-12-04T09:16:47.7764898Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-12-04T09:16:47.7766704Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-12-04T09:16:47.7769550Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-12-04T09:16:47.7771394Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-12-04T09:16:47.7773315Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-12-04T09:16:47.7775867Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-12-04T09:16:47.7777871Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-12-04T09:16:47.7779810Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-12-04T09:16:47.7782102Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-12-04T09:16:47.7784142Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-12-04T09:16:47.7785812Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-12-04T09:16:47.7788249Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-12-04T09:16:47.7790254Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-12-04T09:16:47.7791995Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-12-04T09:16:47.7794327Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-12-04T09:16:47.7796239Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-12-04T09:16:47.7797974Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-12-04T09:16:47.7800636Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-12-04T09:16:47.7802418Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-12-04T09:16:47.7804272Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-12-04T09:16:47.7807023Z * [new branch] gh/tugsbayasgalan/55/base -> origin/gh/tugsbayasgalan/55/base 2025-12-04T09:16:47.7809098Z * [new branch] gh/tugsbayasgalan/55/head -> origin/gh/tugsbayasgalan/55/head 2025-12-04T09:16:47.7811406Z * [new branch] gh/tugsbayasgalan/55/orig -> origin/gh/tugsbayasgalan/55/orig 2025-12-04T09:16:47.7814024Z * [new branch] gh/tugsbayasgalan/59/base -> origin/gh/tugsbayasgalan/59/base 2025-12-04T09:16:47.7815953Z * [new branch] gh/tugsbayasgalan/59/head -> origin/gh/tugsbayasgalan/59/head 2025-12-04T09:16:47.7817753Z * [new branch] gh/tugsbayasgalan/59/orig -> origin/gh/tugsbayasgalan/59/orig 2025-12-04T09:16:47.7820209Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-12-04T09:16:47.7821945Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-12-04T09:16:47.7823775Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-12-04T09:16:47.7826849Z * [new branch] gh/tugsbayasgalan/60/base -> origin/gh/tugsbayasgalan/60/base 2025-12-04T09:16:47.7828866Z * [new branch] gh/tugsbayasgalan/60/head -> origin/gh/tugsbayasgalan/60/head 2025-12-04T09:16:47.7830737Z * [new branch] gh/tugsbayasgalan/60/orig -> origin/gh/tugsbayasgalan/60/orig 2025-12-04T09:16:47.7833859Z * [new branch] gh/tugsbayasgalan/61/base -> origin/gh/tugsbayasgalan/61/base 2025-12-04T09:16:47.7835891Z * [new branch] gh/tugsbayasgalan/61/head -> origin/gh/tugsbayasgalan/61/head 2025-12-04T09:16:47.7840023Z * [new branch] gh/tugsbayasgalan/61/orig -> origin/gh/tugsbayasgalan/61/orig 2025-12-04T09:16:47.7842882Z * [new branch] gh/tugsbayasgalan/63/base -> origin/gh/tugsbayasgalan/63/base 2025-12-04T09:16:47.7844411Z * [new branch] gh/tugsbayasgalan/63/head -> origin/gh/tugsbayasgalan/63/head 2025-12-04T09:16:47.7846249Z * [new branch] gh/tugsbayasgalan/63/orig -> origin/gh/tugsbayasgalan/63/orig 2025-12-04T09:16:47.7849034Z * [new branch] gh/tugsbayasgalan/67/base -> origin/gh/tugsbayasgalan/67/base 2025-12-04T09:16:47.7850917Z * [new branch] gh/tugsbayasgalan/67/head -> origin/gh/tugsbayasgalan/67/head 2025-12-04T09:16:47.7852839Z * [new branch] gh/tugsbayasgalan/67/orig -> origin/gh/tugsbayasgalan/67/orig 2025-12-04T09:16:47.7855619Z * [new branch] gh/tugsbayasgalan/68/base -> origin/gh/tugsbayasgalan/68/base 2025-12-04T09:16:47.7867189Z * [new branch] gh/tugsbayasgalan/68/head -> origin/gh/tugsbayasgalan/68/head 2025-12-04T09:16:47.7867596Z * [new branch] gh/tugsbayasgalan/68/orig -> origin/gh/tugsbayasgalan/68/orig 2025-12-04T09:16:47.7867852Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-12-04T09:16:47.7868088Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-12-04T09:16:47.7868529Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-12-04T09:16:47.7868806Z * [new branch] gh/tugsbayasgalan/70/base -> origin/gh/tugsbayasgalan/70/base 2025-12-04T09:16:47.7871299Z * [new branch] gh/tugsbayasgalan/70/head -> origin/gh/tugsbayasgalan/70/head 2025-12-04T09:16:47.7873406Z * [new branch] gh/tugsbayasgalan/70/orig -> origin/gh/tugsbayasgalan/70/orig 2025-12-04T09:16:47.7875892Z * [new branch] gh/tugsbayasgalan/71/base -> origin/gh/tugsbayasgalan/71/base 2025-12-04T09:16:47.7877793Z * [new branch] gh/tugsbayasgalan/71/head -> origin/gh/tugsbayasgalan/71/head 2025-12-04T09:16:47.7879692Z * [new branch] gh/tugsbayasgalan/71/orig -> origin/gh/tugsbayasgalan/71/orig 2025-12-04T09:16:47.7882414Z * [new branch] gh/tugsbayasgalan/72/base -> origin/gh/tugsbayasgalan/72/base 2025-12-04T09:16:47.7884260Z * [new branch] gh/tugsbayasgalan/72/head -> origin/gh/tugsbayasgalan/72/head 2025-12-04T09:16:47.7886137Z * [new branch] gh/tugsbayasgalan/72/orig -> origin/gh/tugsbayasgalan/72/orig 2025-12-04T09:16:47.7888861Z * [new branch] gh/tugsbayasgalan/73/base -> origin/gh/tugsbayasgalan/73/base 2025-12-04T09:16:47.7890852Z * [new branch] gh/tugsbayasgalan/73/head -> origin/gh/tugsbayasgalan/73/head 2025-12-04T09:16:47.7892858Z * [new branch] gh/tugsbayasgalan/73/orig -> origin/gh/tugsbayasgalan/73/orig 2025-12-04T09:16:47.7895522Z * [new branch] gh/tugsbayasgalan/74/base -> origin/gh/tugsbayasgalan/74/base 2025-12-04T09:16:47.7897959Z * [new branch] gh/tugsbayasgalan/74/head -> origin/gh/tugsbayasgalan/74/head 2025-12-04T09:16:47.7899756Z * [new branch] gh/tugsbayasgalan/74/orig -> origin/gh/tugsbayasgalan/74/orig 2025-12-04T09:16:47.7902365Z * [new branch] gh/tugsbayasgalan/75/base -> origin/gh/tugsbayasgalan/75/base 2025-12-04T09:16:47.7904196Z * [new branch] gh/tugsbayasgalan/75/head -> origin/gh/tugsbayasgalan/75/head 2025-12-04T09:16:47.7906525Z * [new branch] gh/tugsbayasgalan/75/orig -> origin/gh/tugsbayasgalan/75/orig 2025-12-04T09:16:47.7909245Z * [new branch] gh/tugsbayasgalan/76/base -> origin/gh/tugsbayasgalan/76/base 2025-12-04T09:16:47.7911090Z * [new branch] gh/tugsbayasgalan/76/head -> origin/gh/tugsbayasgalan/76/head 2025-12-04T09:16:47.7912880Z * [new branch] gh/tugsbayasgalan/76/orig -> origin/gh/tugsbayasgalan/76/orig 2025-12-04T09:16:47.7915523Z * [new branch] gh/tugsbayasgalan/77/base -> origin/gh/tugsbayasgalan/77/base 2025-12-04T09:16:47.7917358Z * [new branch] gh/tugsbayasgalan/77/head -> origin/gh/tugsbayasgalan/77/head 2025-12-04T09:16:47.7919071Z * [new branch] gh/tugsbayasgalan/77/orig -> origin/gh/tugsbayasgalan/77/orig 2025-12-04T09:16:47.7921682Z * [new branch] gh/tugsbayasgalan/78/base -> origin/gh/tugsbayasgalan/78/base 2025-12-04T09:16:47.7923621Z * [new branch] gh/tugsbayasgalan/78/head -> origin/gh/tugsbayasgalan/78/head 2025-12-04T09:16:47.7925477Z * [new branch] gh/tugsbayasgalan/78/orig -> origin/gh/tugsbayasgalan/78/orig 2025-12-04T09:16:47.7928340Z * [new branch] gh/tugsbayasgalan/79/base -> origin/gh/tugsbayasgalan/79/base 2025-12-04T09:16:47.7930397Z * [new branch] gh/tugsbayasgalan/79/head -> origin/gh/tugsbayasgalan/79/head 2025-12-04T09:16:47.7932167Z * [new branch] gh/tugsbayasgalan/79/orig -> origin/gh/tugsbayasgalan/79/orig 2025-12-04T09:16:47.7934895Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-12-04T09:16:47.7937180Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-12-04T09:16:47.7939225Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-12-04T09:16:47.7941987Z * [new branch] gh/tugsbayasgalan/80/base -> origin/gh/tugsbayasgalan/80/base 2025-12-04T09:16:47.7943788Z * [new branch] gh/tugsbayasgalan/80/head -> origin/gh/tugsbayasgalan/80/head 2025-12-04T09:16:47.7945674Z * [new branch] gh/tugsbayasgalan/80/orig -> origin/gh/tugsbayasgalan/80/orig 2025-12-04T09:16:47.7948350Z * [new branch] gh/tugsbayasgalan/81/base -> origin/gh/tugsbayasgalan/81/base 2025-12-04T09:16:47.7950152Z * [new branch] gh/tugsbayasgalan/81/head -> origin/gh/tugsbayasgalan/81/head 2025-12-04T09:16:47.7952059Z * [new branch] gh/tugsbayasgalan/81/orig -> origin/gh/tugsbayasgalan/81/orig 2025-12-04T09:16:47.7955262Z * [new branch] gh/tugsbayasgalan/82/base -> origin/gh/tugsbayasgalan/82/base 2025-12-04T09:16:47.7957266Z * [new branch] gh/tugsbayasgalan/82/head -> origin/gh/tugsbayasgalan/82/head 2025-12-04T09:16:47.7959151Z * [new branch] gh/tugsbayasgalan/82/orig -> origin/gh/tugsbayasgalan/82/orig 2025-12-04T09:16:47.7961721Z * [new branch] gh/tugsbayasgalan/83/base -> origin/gh/tugsbayasgalan/83/base 2025-12-04T09:16:47.7963623Z * [new branch] gh/tugsbayasgalan/83/head -> origin/gh/tugsbayasgalan/83/head 2025-12-04T09:16:47.7965487Z * [new branch] gh/tugsbayasgalan/83/orig -> origin/gh/tugsbayasgalan/83/orig 2025-12-04T09:16:47.7968051Z * [new branch] gh/tugsbayasgalan/84/base -> origin/gh/tugsbayasgalan/84/base 2025-12-04T09:16:47.7969978Z * [new branch] gh/tugsbayasgalan/84/head -> origin/gh/tugsbayasgalan/84/head 2025-12-04T09:16:47.7971774Z * [new branch] gh/tugsbayasgalan/84/orig -> origin/gh/tugsbayasgalan/84/orig 2025-12-04T09:16:47.7974236Z * [new branch] gh/tugsbayasgalan/85/base -> origin/gh/tugsbayasgalan/85/base 2025-12-04T09:16:47.7976171Z * [new branch] gh/tugsbayasgalan/85/head -> origin/gh/tugsbayasgalan/85/head 2025-12-04T09:16:47.7977981Z * [new branch] gh/tugsbayasgalan/85/orig -> origin/gh/tugsbayasgalan/85/orig 2025-12-04T09:16:47.7980563Z * [new branch] gh/tugsbayasgalan/86/base -> origin/gh/tugsbayasgalan/86/base 2025-12-04T09:16:47.7982454Z * [new branch] gh/tugsbayasgalan/86/head -> origin/gh/tugsbayasgalan/86/head 2025-12-04T09:16:47.7984204Z * [new branch] gh/tugsbayasgalan/86/orig -> origin/gh/tugsbayasgalan/86/orig 2025-12-04T09:16:47.7987324Z * [new branch] gh/tugsbayasgalan/87/base -> origin/gh/tugsbayasgalan/87/base 2025-12-04T09:16:47.7989195Z * [new branch] gh/tugsbayasgalan/87/head -> origin/gh/tugsbayasgalan/87/head 2025-12-04T09:16:47.7991019Z * [new branch] gh/tugsbayasgalan/87/orig -> origin/gh/tugsbayasgalan/87/orig 2025-12-04T09:16:47.7993680Z * [new branch] gh/tugsbayasgalan/88/base -> origin/gh/tugsbayasgalan/88/base 2025-12-04T09:16:47.7995506Z * [new branch] gh/tugsbayasgalan/88/head -> origin/gh/tugsbayasgalan/88/head 2025-12-04T09:16:47.7997394Z * [new branch] gh/tugsbayasgalan/88/orig -> origin/gh/tugsbayasgalan/88/orig 2025-12-04T09:16:47.7999976Z * [new branch] gh/tugsbayasgalan/89/base -> origin/gh/tugsbayasgalan/89/base 2025-12-04T09:16:47.8001766Z * [new branch] gh/tugsbayasgalan/89/head -> origin/gh/tugsbayasgalan/89/head 2025-12-04T09:16:47.8003660Z * [new branch] gh/tugsbayasgalan/89/orig -> origin/gh/tugsbayasgalan/89/orig 2025-12-04T09:16:47.8006253Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-12-04T09:16:47.8008017Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-12-04T09:16:47.8009923Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-12-04T09:16:47.8012851Z * [new branch] gh/tugsbayasgalan/90/base -> origin/gh/tugsbayasgalan/90/base 2025-12-04T09:16:47.8014646Z * [new branch] gh/tugsbayasgalan/90/head -> origin/gh/tugsbayasgalan/90/head 2025-12-04T09:16:47.8016396Z * [new branch] gh/tugsbayasgalan/90/orig -> origin/gh/tugsbayasgalan/90/orig 2025-12-04T09:16:47.8019815Z * [new branch] gh/tugsbayasgalan/91/base -> origin/gh/tugsbayasgalan/91/base 2025-12-04T09:16:47.8021348Z * [new branch] gh/tugsbayasgalan/91/head -> origin/gh/tugsbayasgalan/91/head 2025-12-04T09:16:47.8023093Z * [new branch] gh/tugsbayasgalan/91/orig -> origin/gh/tugsbayasgalan/91/orig 2025-12-04T09:16:47.8025703Z * [new branch] gh/tugsbayasgalan/92/base -> origin/gh/tugsbayasgalan/92/base 2025-12-04T09:16:47.8027678Z * [new branch] gh/tugsbayasgalan/92/head -> origin/gh/tugsbayasgalan/92/head 2025-12-04T09:16:47.8029561Z * [new branch] gh/tugsbayasgalan/92/orig -> origin/gh/tugsbayasgalan/92/orig 2025-12-04T09:16:47.8032215Z * [new branch] gh/tugsbayasgalan/93/base -> origin/gh/tugsbayasgalan/93/base 2025-12-04T09:16:47.8034106Z * [new branch] gh/tugsbayasgalan/93/head -> origin/gh/tugsbayasgalan/93/head 2025-12-04T09:16:47.8036121Z * [new branch] gh/tugsbayasgalan/93/orig -> origin/gh/tugsbayasgalan/93/orig 2025-12-04T09:16:47.8039491Z * [new branch] gh/v0i0/14/base -> origin/gh/v0i0/14/base 2025-12-04T09:16:47.8041184Z * [new branch] gh/v0i0/14/head -> origin/gh/v0i0/14/head 2025-12-04T09:16:47.8042921Z * [new branch] gh/v0i0/14/orig -> origin/gh/v0i0/14/orig 2025-12-04T09:16:47.8045282Z * [new branch] gh/v0i0/15/base -> origin/gh/v0i0/15/base 2025-12-04T09:16:47.8047201Z * [new branch] gh/v0i0/15/head -> origin/gh/v0i0/15/head 2025-12-04T09:16:47.8049239Z * [new branch] gh/v0i0/15/orig -> origin/gh/v0i0/15/orig 2025-12-04T09:16:47.8051755Z * [new branch] gh/v0i0/16/base -> origin/gh/v0i0/16/base 2025-12-04T09:16:47.8053540Z * [new branch] gh/v0i0/16/head -> origin/gh/v0i0/16/head 2025-12-04T09:16:47.8055377Z * [new branch] gh/v0i0/16/orig -> origin/gh/v0i0/16/orig 2025-12-04T09:16:47.8057833Z * [new branch] gh/v0i0/17/base -> origin/gh/v0i0/17/base 2025-12-04T09:16:47.8059697Z * [new branch] gh/v0i0/17/head -> origin/gh/v0i0/17/head 2025-12-04T09:16:47.8061585Z * [new branch] gh/v0i0/17/orig -> origin/gh/v0i0/17/orig 2025-12-04T09:16:47.8064230Z * [new branch] gh/v0i0/18/base -> origin/gh/v0i0/18/base 2025-12-04T09:16:47.8066105Z * [new branch] gh/v0i0/18/head -> origin/gh/v0i0/18/head 2025-12-04T09:16:47.8067948Z * [new branch] gh/v0i0/18/orig -> origin/gh/v0i0/18/orig 2025-12-04T09:16:47.8070447Z * [new branch] gh/v0i0/19/base -> origin/gh/v0i0/19/base 2025-12-04T09:16:47.8072278Z * [new branch] gh/v0i0/19/head -> origin/gh/v0i0/19/head 2025-12-04T09:16:47.8074081Z * [new branch] gh/v0i0/19/orig -> origin/gh/v0i0/19/orig 2025-12-04T09:16:47.8077225Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-12-04T09:16:47.8079058Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-12-04T09:16:47.8081513Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-12-04T09:16:47.8083648Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-12-04T09:16:47.8085459Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-12-04T09:16:47.8088301Z * [new branch] gh/vishal9-team/3/base -> origin/gh/vishal9-team/3/base 2025-12-04T09:16:47.8090563Z * [new branch] gh/vishal9-team/3/head -> origin/gh/vishal9-team/3/head 2025-12-04T09:16:47.8092415Z * [new branch] gh/vishal9-team/3/orig -> origin/gh/vishal9-team/3/orig 2025-12-04T09:16:47.8095395Z * [new branch] gh/vishal9-team/4/base -> origin/gh/vishal9-team/4/base 2025-12-04T09:16:47.8097125Z * [new branch] gh/vishal9-team/4/head -> origin/gh/vishal9-team/4/head 2025-12-04T09:16:47.8098967Z * [new branch] gh/vishal9-team/4/orig -> origin/gh/vishal9-team/4/orig 2025-12-04T09:16:47.8101881Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-12-04T09:16:47.8104394Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-12-04T09:16:47.8106839Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-12-04T09:16:47.8110046Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-12-04T09:16:47.8112078Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-12-04T09:16:47.8113931Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-12-04T09:16:47.8116462Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-12-04T09:16:47.8118352Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-12-04T09:16:47.8120330Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-12-04T09:16:47.8122768Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-12-04T09:16:47.8124657Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-12-04T09:16:47.8126459Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-12-04T09:16:47.8129324Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-12-04T09:16:47.8131323Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-12-04T09:16:47.8132781Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-12-04T09:16:47.8135480Z * [new branch] gh/wconstab/448/base -> origin/gh/wconstab/448/base 2025-12-04T09:16:47.8137573Z * [new branch] gh/wconstab/448/head -> origin/gh/wconstab/448/head 2025-12-04T09:16:47.8139532Z * [new branch] gh/wconstab/448/orig -> origin/gh/wconstab/448/orig 2025-12-04T09:16:47.8142002Z * [new branch] gh/wconstab/449/base -> origin/gh/wconstab/449/base 2025-12-04T09:16:47.8143783Z * [new branch] gh/wconstab/449/head -> origin/gh/wconstab/449/head 2025-12-04T09:16:47.8145638Z * [new branch] gh/wconstab/449/orig -> origin/gh/wconstab/449/orig 2025-12-04T09:16:47.8147961Z * [new branch] gh/wconstab/450/base -> origin/gh/wconstab/450/base 2025-12-04T09:16:47.8149895Z * [new branch] gh/wconstab/450/head -> origin/gh/wconstab/450/head 2025-12-04T09:16:47.8151689Z * [new branch] gh/wconstab/450/orig -> origin/gh/wconstab/450/orig 2025-12-04T09:16:47.8153995Z * [new branch] gh/wconstab/451/base -> origin/gh/wconstab/451/base 2025-12-04T09:16:47.8155963Z * [new branch] gh/wconstab/451/head -> origin/gh/wconstab/451/head 2025-12-04T09:16:47.8157855Z * [new branch] gh/wconstab/451/orig -> origin/gh/wconstab/451/orig 2025-12-04T09:16:47.8160474Z * [new branch] gh/wconstab/452/base -> origin/gh/wconstab/452/base 2025-12-04T09:16:47.8162193Z * [new branch] gh/wconstab/452/head -> origin/gh/wconstab/452/head 2025-12-04T09:16:47.8164134Z * [new branch] gh/wconstab/452/orig -> origin/gh/wconstab/452/orig 2025-12-04T09:16:47.8166377Z * [new branch] gh/wconstab/453/base -> origin/gh/wconstab/453/base 2025-12-04T09:16:47.8168436Z * [new branch] gh/wconstab/453/head -> origin/gh/wconstab/453/head 2025-12-04T09:16:47.8170430Z * [new branch] gh/wconstab/453/orig -> origin/gh/wconstab/453/orig 2025-12-04T09:16:47.8173479Z * [new branch] gh/wconstab/454/base -> origin/gh/wconstab/454/base 2025-12-04T09:16:47.8175573Z * [new branch] gh/wconstab/454/head -> origin/gh/wconstab/454/head 2025-12-04T09:16:47.8176379Z * [new branch] gh/wconstab/454/orig -> origin/gh/wconstab/454/orig 2025-12-04T09:16:47.8179248Z * [new branch] gh/wconstab/455/base -> origin/gh/wconstab/455/base 2025-12-04T09:16:47.8181124Z * [new branch] gh/wconstab/455/head -> origin/gh/wconstab/455/head 2025-12-04T09:16:47.8182973Z * [new branch] gh/wconstab/455/orig -> origin/gh/wconstab/455/orig 2025-12-04T09:16:47.8185835Z * [new branch] gh/wconstab/456/base -> origin/gh/wconstab/456/base 2025-12-04T09:16:47.8187968Z * [new branch] gh/wconstab/456/head -> origin/gh/wconstab/456/head 2025-12-04T09:16:47.8189864Z * [new branch] gh/wconstab/456/orig -> origin/gh/wconstab/456/orig 2025-12-04T09:16:47.8192337Z * [new branch] gh/wconstab/457/base -> origin/gh/wconstab/457/base 2025-12-04T09:16:47.8194218Z * [new branch] gh/wconstab/457/head -> origin/gh/wconstab/457/head 2025-12-04T09:16:47.8196013Z * [new branch] gh/wconstab/457/orig -> origin/gh/wconstab/457/orig 2025-12-04T09:16:47.8198588Z * [new branch] gh/wconstab/458/base -> origin/gh/wconstab/458/base 2025-12-04T09:16:47.8200495Z * [new branch] gh/wconstab/458/head -> origin/gh/wconstab/458/head 2025-12-04T09:16:47.8202231Z * [new branch] gh/wconstab/458/orig -> origin/gh/wconstab/458/orig 2025-12-04T09:16:47.8204656Z * [new branch] gh/wconstab/459/base -> origin/gh/wconstab/459/base 2025-12-04T09:16:47.8206528Z * [new branch] gh/wconstab/459/head -> origin/gh/wconstab/459/head 2025-12-04T09:16:47.8208427Z * [new branch] gh/wconstab/459/orig -> origin/gh/wconstab/459/orig 2025-12-04T09:16:47.8211907Z * [new branch] gh/wconstab/460/base -> origin/gh/wconstab/460/base 2025-12-04T09:16:47.8213915Z * [new branch] gh/wconstab/460/head -> origin/gh/wconstab/460/head 2025-12-04T09:16:47.8215794Z * [new branch] gh/wconstab/460/orig -> origin/gh/wconstab/460/orig 2025-12-04T09:16:47.8218442Z * [new branch] gh/wconstab/461/base -> origin/gh/wconstab/461/base 2025-12-04T09:16:47.8220271Z * [new branch] gh/wconstab/461/head -> origin/gh/wconstab/461/head 2025-12-04T09:16:47.8222062Z * [new branch] gh/wconstab/461/orig -> origin/gh/wconstab/461/orig 2025-12-04T09:16:47.8224445Z * [new branch] gh/wconstab/462/base -> origin/gh/wconstab/462/base 2025-12-04T09:16:47.8226393Z * [new branch] gh/wconstab/462/head -> origin/gh/wconstab/462/head 2025-12-04T09:16:47.8228274Z * [new branch] gh/wconstab/462/orig -> origin/gh/wconstab/462/orig 2025-12-04T09:16:47.8231434Z * [new branch] gh/wconstab/463/base -> origin/gh/wconstab/463/base 2025-12-04T09:16:47.8233304Z * [new branch] gh/wconstab/463/head -> origin/gh/wconstab/463/head 2025-12-04T09:16:47.8235141Z * [new branch] gh/wconstab/463/orig -> origin/gh/wconstab/463/orig 2025-12-04T09:16:47.8241012Z * [new branch] gh/wconstab/464/base -> origin/gh/wconstab/464/base 2025-12-04T09:16:47.8243120Z * [new branch] gh/wconstab/464/head -> origin/gh/wconstab/464/head 2025-12-04T09:16:47.8244851Z * [new branch] gh/wconstab/464/orig -> origin/gh/wconstab/464/orig 2025-12-04T09:16:47.8247283Z * [new branch] gh/wconstab/465/base -> origin/gh/wconstab/465/base 2025-12-04T09:16:47.8249383Z * [new branch] gh/wconstab/465/head -> origin/gh/wconstab/465/head 2025-12-04T09:16:47.8251050Z * [new branch] gh/wconstab/465/orig -> origin/gh/wconstab/465/orig 2025-12-04T09:16:47.8253879Z * [new branch] gh/wconstab/466/base -> origin/gh/wconstab/466/base 2025-12-04T09:16:47.8255487Z * [new branch] gh/wconstab/466/head -> origin/gh/wconstab/466/head 2025-12-04T09:16:47.8257254Z * [new branch] gh/wconstab/466/orig -> origin/gh/wconstab/466/orig 2025-12-04T09:16:47.8260108Z * [new branch] gh/wconstab/467/base -> origin/gh/wconstab/467/base 2025-12-04T09:16:47.8262024Z * [new branch] gh/wconstab/467/head -> origin/gh/wconstab/467/head 2025-12-04T09:16:47.8264285Z * [new branch] gh/wconstab/467/orig -> origin/gh/wconstab/467/orig 2025-12-04T09:16:47.8266875Z * [new branch] gh/wconstab/468/base -> origin/gh/wconstab/468/base 2025-12-04T09:16:47.8268684Z * [new branch] gh/wconstab/468/head -> origin/gh/wconstab/468/head 2025-12-04T09:16:47.8270539Z * [new branch] gh/wconstab/468/orig -> origin/gh/wconstab/468/orig 2025-12-04T09:16:47.8273748Z * [new branch] gh/weifengpy/39/base -> origin/gh/weifengpy/39/base 2025-12-04T09:16:47.8275941Z * [new branch] gh/weifengpy/39/head -> origin/gh/weifengpy/39/head 2025-12-04T09:16:47.8277472Z * [new branch] gh/weifengpy/39/orig -> origin/gh/weifengpy/39/orig 2025-12-04T09:16:47.8280076Z * [new branch] gh/weifengpy/40/base -> origin/gh/weifengpy/40/base 2025-12-04T09:16:47.8281842Z * [new branch] gh/weifengpy/40/head -> origin/gh/weifengpy/40/head 2025-12-04T09:16:47.8283670Z * [new branch] gh/weifengpy/40/orig -> origin/gh/weifengpy/40/orig 2025-12-04T09:16:47.8286311Z * [new branch] gh/weifengpy/41/base -> origin/gh/weifengpy/41/base 2025-12-04T09:16:47.8288640Z * [new branch] gh/weifengpy/41/head -> origin/gh/weifengpy/41/head 2025-12-04T09:16:47.8290500Z * [new branch] gh/weifengpy/41/orig -> origin/gh/weifengpy/41/orig 2025-12-04T09:16:47.8293652Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-12-04T09:16:47.8295479Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-12-04T09:16:47.8297318Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-12-04T09:16:47.8300041Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-12-04T09:16:47.8302096Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-12-04T09:16:47.8303865Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-12-04T09:16:47.8306372Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-12-04T09:16:47.8308162Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-12-04T09:16:47.8310051Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-12-04T09:16:47.8312605Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-12-04T09:16:47.8314614Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-12-04T09:16:47.8316482Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-12-04T09:16:47.8319203Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-12-04T09:16:47.8320841Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-12-04T09:16:47.8322603Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-12-04T09:16:47.8325328Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-12-04T09:16:47.8327307Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-12-04T09:16:47.8329371Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-12-04T09:16:47.8331734Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-12-04T09:16:47.8333596Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-12-04T09:16:47.8335727Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-12-04T09:16:47.8338445Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-12-04T09:16:47.8340466Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-12-04T09:16:47.8342343Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-12-04T09:16:47.8344940Z * [new branch] gh/williamwen42/309/base -> origin/gh/williamwen42/309/base 2025-12-04T09:16:47.8346841Z * [new branch] gh/williamwen42/309/head -> origin/gh/williamwen42/309/head 2025-12-04T09:16:47.8348716Z * [new branch] gh/williamwen42/309/orig -> origin/gh/williamwen42/309/orig 2025-12-04T09:16:47.8351189Z * [new branch] gh/williamwen42/310/base -> origin/gh/williamwen42/310/base 2025-12-04T09:16:47.8353058Z * [new branch] gh/williamwen42/310/head -> origin/gh/williamwen42/310/head 2025-12-04T09:16:47.8355018Z * [new branch] gh/williamwen42/310/orig -> origin/gh/williamwen42/310/orig 2025-12-04T09:16:47.8358650Z * [new branch] gh/williamwen42/311/base -> origin/gh/williamwen42/311/base 2025-12-04T09:16:47.8360488Z * [new branch] gh/williamwen42/311/head -> origin/gh/williamwen42/311/head 2025-12-04T09:16:47.8362254Z * [new branch] gh/williamwen42/311/orig -> origin/gh/williamwen42/311/orig 2025-12-04T09:16:47.8364627Z * [new branch] gh/williamwen42/319/base -> origin/gh/williamwen42/319/base 2025-12-04T09:16:47.8366610Z * [new branch] gh/williamwen42/319/head -> origin/gh/williamwen42/319/head 2025-12-04T09:16:47.8368567Z * [new branch] gh/williamwen42/319/orig -> origin/gh/williamwen42/319/orig 2025-12-04T09:16:47.8371202Z * [new branch] gh/williamwen42/325/base -> origin/gh/williamwen42/325/base 2025-12-04T09:16:47.8373067Z * [new branch] gh/williamwen42/325/head -> origin/gh/williamwen42/325/head 2025-12-04T09:16:47.8374852Z * [new branch] gh/williamwen42/325/orig -> origin/gh/williamwen42/325/orig 2025-12-04T09:16:47.8377358Z * [new branch] gh/williamwen42/326/base -> origin/gh/williamwen42/326/base 2025-12-04T09:16:47.8379580Z * [new branch] gh/williamwen42/326/head -> origin/gh/williamwen42/326/head 2025-12-04T09:16:47.8381259Z * [new branch] gh/williamwen42/326/orig -> origin/gh/williamwen42/326/orig 2025-12-04T09:16:47.8383772Z * [new branch] gh/williamwen42/327/base -> origin/gh/williamwen42/327/base 2025-12-04T09:16:47.8385607Z * [new branch] gh/williamwen42/327/head -> origin/gh/williamwen42/327/head 2025-12-04T09:16:47.8387377Z * [new branch] gh/williamwen42/327/orig -> origin/gh/williamwen42/327/orig 2025-12-04T09:16:47.8390052Z * [new branch] gh/williamwen42/328/base -> origin/gh/williamwen42/328/base 2025-12-04T09:16:47.8392258Z * [new branch] gh/williamwen42/328/head -> origin/gh/williamwen42/328/head 2025-12-04T09:16:47.8393976Z * [new branch] gh/williamwen42/328/orig -> origin/gh/williamwen42/328/orig 2025-12-04T09:16:47.8397100Z * [new branch] gh/williamwen42/329/base -> origin/gh/williamwen42/329/base 2025-12-04T09:16:47.8399099Z * [new branch] gh/williamwen42/329/head -> origin/gh/williamwen42/329/head 2025-12-04T09:16:47.8400862Z * [new branch] gh/williamwen42/329/orig -> origin/gh/williamwen42/329/orig 2025-12-04T09:16:47.8403432Z * [new branch] gh/williamwen42/330/base -> origin/gh/williamwen42/330/base 2025-12-04T09:16:47.8405318Z * [new branch] gh/williamwen42/330/head -> origin/gh/williamwen42/330/head 2025-12-04T09:16:47.8407156Z * [new branch] gh/williamwen42/330/orig -> origin/gh/williamwen42/330/orig 2025-12-04T09:16:47.8409822Z * [new branch] gh/williamwen42/331/base -> origin/gh/williamwen42/331/base 2025-12-04T09:16:47.8411768Z * [new branch] gh/williamwen42/331/head -> origin/gh/williamwen42/331/head 2025-12-04T09:16:47.8413513Z * [new branch] gh/williamwen42/331/orig -> origin/gh/williamwen42/331/orig 2025-12-04T09:16:47.8416026Z * [new branch] gh/williamwen42/332/base -> origin/gh/williamwen42/332/base 2025-12-04T09:16:47.8418289Z * [new branch] gh/williamwen42/332/head -> origin/gh/williamwen42/332/head 2025-12-04T09:16:47.8420239Z * [new branch] gh/williamwen42/332/orig -> origin/gh/williamwen42/332/orig 2025-12-04T09:16:47.8422957Z * [new branch] gh/williamwen42/333/base -> origin/gh/williamwen42/333/base 2025-12-04T09:16:47.8424792Z * [new branch] gh/williamwen42/333/head -> origin/gh/williamwen42/333/head 2025-12-04T09:16:47.8426648Z * [new branch] gh/williamwen42/333/orig -> origin/gh/williamwen42/333/orig 2025-12-04T09:16:47.8429279Z * [new branch] gh/williamwen42/334/base -> origin/gh/williamwen42/334/base 2025-12-04T09:16:47.8431070Z * [new branch] gh/williamwen42/334/head -> origin/gh/williamwen42/334/head 2025-12-04T09:16:47.8432899Z * [new branch] gh/williamwen42/334/orig -> origin/gh/williamwen42/334/orig 2025-12-04T09:16:47.8439119Z * [new branch] gh/williamwen42/335/base -> origin/gh/williamwen42/335/base 2025-12-04T09:16:47.8440925Z * [new branch] gh/williamwen42/335/head -> origin/gh/williamwen42/335/head 2025-12-04T09:16:47.8442688Z * [new branch] gh/williamwen42/335/orig -> origin/gh/williamwen42/335/orig 2025-12-04T09:16:47.8445306Z * [new branch] gh/williamwen42/336/base -> origin/gh/williamwen42/336/base 2025-12-04T09:16:47.8447170Z * [new branch] gh/williamwen42/336/head -> origin/gh/williamwen42/336/head 2025-12-04T09:16:47.8449119Z * [new branch] gh/williamwen42/336/orig -> origin/gh/williamwen42/336/orig 2025-12-04T09:16:47.8451693Z * [new branch] gh/williamwen42/337/base -> origin/gh/williamwen42/337/base 2025-12-04T09:16:47.8453523Z * [new branch] gh/williamwen42/337/head -> origin/gh/williamwen42/337/head 2025-12-04T09:16:47.8455346Z * [new branch] gh/williamwen42/337/orig -> origin/gh/williamwen42/337/orig 2025-12-04T09:16:47.8457984Z * [new branch] gh/williamwen42/338/base -> origin/gh/williamwen42/338/base 2025-12-04T09:16:47.8459807Z * [new branch] gh/williamwen42/338/head -> origin/gh/williamwen42/338/head 2025-12-04T09:16:47.8461598Z * [new branch] gh/williamwen42/338/orig -> origin/gh/williamwen42/338/orig 2025-12-04T09:16:47.8464178Z * [new branch] gh/williamwen42/339/base -> origin/gh/williamwen42/339/base 2025-12-04T09:16:47.8466177Z * [new branch] gh/williamwen42/339/head -> origin/gh/williamwen42/339/head 2025-12-04T09:16:47.8467853Z * [new branch] gh/williamwen42/339/orig -> origin/gh/williamwen42/339/orig 2025-12-04T09:16:47.8470546Z * [new branch] gh/williamwen42/340/base -> origin/gh/williamwen42/340/base 2025-12-04T09:16:47.8472441Z * [new branch] gh/williamwen42/340/head -> origin/gh/williamwen42/340/head 2025-12-04T09:16:47.8474187Z * [new branch] gh/williamwen42/340/orig -> origin/gh/williamwen42/340/orig 2025-12-04T09:16:47.8476838Z * [new branch] gh/williamwen42/341/base -> origin/gh/williamwen42/341/base 2025-12-04T09:16:47.8479360Z * [new branch] gh/williamwen42/341/head -> origin/gh/williamwen42/341/head 2025-12-04T09:16:47.8481153Z * [new branch] gh/williamwen42/341/orig -> origin/gh/williamwen42/341/orig 2025-12-04T09:16:47.8483733Z * [new branch] gh/williamwen42/342/base -> origin/gh/williamwen42/342/base 2025-12-04T09:16:47.8485551Z * [new branch] gh/williamwen42/342/head -> origin/gh/williamwen42/342/head 2025-12-04T09:16:47.8487344Z * [new branch] gh/williamwen42/342/orig -> origin/gh/williamwen42/342/orig 2025-12-04T09:16:47.8490137Z * [new branch] gh/williamwen42/343/base -> origin/gh/williamwen42/343/base 2025-12-04T09:16:47.8491973Z * [new branch] gh/williamwen42/343/head -> origin/gh/williamwen42/343/head 2025-12-04T09:16:47.8493768Z * [new branch] gh/williamwen42/343/orig -> origin/gh/williamwen42/343/orig 2025-12-04T09:16:47.8496369Z * [new branch] gh/williamwen42/344/base -> origin/gh/williamwen42/344/base 2025-12-04T09:16:47.8498394Z * [new branch] gh/williamwen42/344/head -> origin/gh/williamwen42/344/head 2025-12-04T09:16:47.8500153Z * [new branch] gh/williamwen42/344/orig -> origin/gh/williamwen42/344/orig 2025-12-04T09:16:47.8502775Z * [new branch] gh/williamwen42/345/base -> origin/gh/williamwen42/345/base 2025-12-04T09:16:47.8504596Z * [new branch] gh/williamwen42/345/head -> origin/gh/williamwen42/345/head 2025-12-04T09:16:47.8506479Z * [new branch] gh/williamwen42/345/orig -> origin/gh/williamwen42/345/orig 2025-12-04T09:16:47.8509073Z * [new branch] gh/williamwen42/346/base -> origin/gh/williamwen42/346/base 2025-12-04T09:16:47.8510929Z * [new branch] gh/williamwen42/346/head -> origin/gh/williamwen42/346/head 2025-12-04T09:16:47.8512836Z * [new branch] gh/williamwen42/346/orig -> origin/gh/williamwen42/346/orig 2025-12-04T09:16:47.8515540Z * [new branch] gh/williamwen42/347/base -> origin/gh/williamwen42/347/base 2025-12-04T09:16:47.8517248Z * [new branch] gh/williamwen42/347/head -> origin/gh/williamwen42/347/head 2025-12-04T09:16:47.8519088Z * [new branch] gh/williamwen42/347/orig -> origin/gh/williamwen42/347/orig 2025-12-04T09:16:47.8521550Z * [new branch] gh/williamwen42/348/base -> origin/gh/williamwen42/348/base 2025-12-04T09:16:47.8523430Z * [new branch] gh/williamwen42/348/head -> origin/gh/williamwen42/348/head 2025-12-04T09:16:47.8525161Z * [new branch] gh/williamwen42/348/orig -> origin/gh/williamwen42/348/orig 2025-12-04T09:16:47.8527655Z * [new branch] gh/williamwen42/349/base -> origin/gh/williamwen42/349/base 2025-12-04T09:16:47.8529591Z * [new branch] gh/williamwen42/349/head -> origin/gh/williamwen42/349/head 2025-12-04T09:16:47.8531388Z * [new branch] gh/williamwen42/349/orig -> origin/gh/williamwen42/349/orig 2025-12-04T09:16:47.8533956Z * [new branch] gh/williamwen42/350/base -> origin/gh/williamwen42/350/base 2025-12-04T09:16:47.8536051Z * [new branch] gh/williamwen42/350/head -> origin/gh/williamwen42/350/head 2025-12-04T09:16:47.8538141Z * [new branch] gh/williamwen42/350/orig -> origin/gh/williamwen42/350/orig 2025-12-04T09:16:47.8540708Z * [new branch] gh/williamwen42/351/base -> origin/gh/williamwen42/351/base 2025-12-04T09:16:47.8542716Z * [new branch] gh/williamwen42/351/head -> origin/gh/williamwen42/351/head 2025-12-04T09:16:47.8544735Z * [new branch] gh/williamwen42/351/orig -> origin/gh/williamwen42/351/orig 2025-12-04T09:16:47.8547316Z * [new branch] gh/williamwen42/352/base -> origin/gh/williamwen42/352/base 2025-12-04T09:16:47.8549332Z * [new branch] gh/williamwen42/352/head -> origin/gh/williamwen42/352/head 2025-12-04T09:16:47.8551114Z * [new branch] gh/williamwen42/352/orig -> origin/gh/williamwen42/352/orig 2025-12-04T09:16:47.8553848Z * [new branch] gh/williamwen42/353/base -> origin/gh/williamwen42/353/base 2025-12-04T09:16:47.8555962Z * [new branch] gh/williamwen42/353/head -> origin/gh/williamwen42/353/head 2025-12-04T09:16:47.8558295Z * [new branch] gh/williamwen42/353/orig -> origin/gh/williamwen42/353/orig 2025-12-04T09:16:47.8560856Z * [new branch] gh/williamwen42/354/base -> origin/gh/williamwen42/354/base 2025-12-04T09:16:47.8562470Z * [new branch] gh/williamwen42/354/head -> origin/gh/williamwen42/354/head 2025-12-04T09:16:47.8564297Z * [new branch] gh/williamwen42/354/orig -> origin/gh/williamwen42/354/orig 2025-12-04T09:16:47.8568538Z * [new branch] gh/williamwen42/355/base -> origin/gh/williamwen42/355/base 2025-12-04T09:16:47.8569479Z * [new branch] gh/williamwen42/355/head -> origin/gh/williamwen42/355/head 2025-12-04T09:16:47.8571269Z * [new branch] gh/williamwen42/355/orig -> origin/gh/williamwen42/355/orig 2025-12-04T09:16:47.8574202Z * [new branch] gh/williamwen42/356/base -> origin/gh/williamwen42/356/base 2025-12-04T09:16:47.8575851Z * [new branch] gh/williamwen42/356/head -> origin/gh/williamwen42/356/head 2025-12-04T09:16:47.8577643Z * [new branch] gh/williamwen42/356/orig -> origin/gh/williamwen42/356/orig 2025-12-04T09:16:47.8581007Z * [new branch] gh/williamwen42/357/base -> origin/gh/williamwen42/357/base 2025-12-04T09:16:47.8582164Z * [new branch] gh/williamwen42/357/head -> origin/gh/williamwen42/357/head 2025-12-04T09:16:47.8583947Z * [new branch] gh/williamwen42/357/orig -> origin/gh/williamwen42/357/orig 2025-12-04T09:16:47.8586853Z * [new branch] gh/williamwen42/358/base -> origin/gh/williamwen42/358/base 2025-12-04T09:16:47.8588275Z * [new branch] gh/williamwen42/358/head -> origin/gh/williamwen42/358/head 2025-12-04T09:16:47.8590659Z * [new branch] gh/williamwen42/358/orig -> origin/gh/williamwen42/358/orig 2025-12-04T09:16:47.8593509Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-12-04T09:16:47.8595085Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-12-04T09:16:47.8598291Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-12-04T09:16:47.8599945Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-12-04T09:16:47.8602703Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-12-04T09:16:47.8604223Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-12-04T09:16:47.8606394Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-12-04T09:16:47.8609102Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-12-04T09:16:47.8610653Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-12-04T09:16:47.8612413Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-12-04T09:16:47.8615456Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-12-04T09:16:47.8616737Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-12-04T09:16:47.8618556Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-12-04T09:16:47.8621356Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-12-04T09:16:47.8622877Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-12-04T09:16:47.8624863Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-12-04T09:16:47.8627477Z * [new branch] gh/xmfan/309/base -> origin/gh/xmfan/309/base 2025-12-04T09:16:47.8628883Z * [new branch] gh/xmfan/309/head -> origin/gh/xmfan/309/head 2025-12-04T09:16:47.8630896Z * [new branch] gh/xmfan/309/orig -> origin/gh/xmfan/309/orig 2025-12-04T09:16:47.8633450Z * [new branch] gh/xmfan/310/base -> origin/gh/xmfan/310/base 2025-12-04T09:16:47.8635004Z * [new branch] gh/xmfan/310/head -> origin/gh/xmfan/310/head 2025-12-04T09:16:47.8639657Z * [new branch] gh/xmfan/310/orig -> origin/gh/xmfan/310/orig 2025-12-04T09:16:47.8642095Z * [new branch] gh/xmfan/311/base -> origin/gh/xmfan/311/base 2025-12-04T09:16:47.8643632Z * [new branch] gh/xmfan/311/head -> origin/gh/xmfan/311/head 2025-12-04T09:16:47.8645392Z * [new branch] gh/xmfan/311/orig -> origin/gh/xmfan/311/orig 2025-12-04T09:16:47.8648233Z * [new branch] gh/xmfan/312/base -> origin/gh/xmfan/312/base 2025-12-04T09:16:47.8649855Z * [new branch] gh/xmfan/312/head -> origin/gh/xmfan/312/head 2025-12-04T09:16:47.8651928Z * [new branch] gh/xmfan/312/orig -> origin/gh/xmfan/312/orig 2025-12-04T09:16:47.8654642Z * [new branch] gh/xmfan/313/base -> origin/gh/xmfan/313/base 2025-12-04T09:16:47.8656121Z * [new branch] gh/xmfan/313/head -> origin/gh/xmfan/313/head 2025-12-04T09:16:47.8658258Z * [new branch] gh/xmfan/313/orig -> origin/gh/xmfan/313/orig 2025-12-04T09:16:47.8661398Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-12-04T09:16:47.8662993Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-12-04T09:16:47.8665032Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-12-04T09:16:47.8667815Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-12-04T09:16:47.8670168Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-12-04T09:16:47.8671740Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-12-04T09:16:47.8674712Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-12-04T09:16:47.8676355Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-12-04T09:16:47.8678060Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-12-04T09:16:47.8681254Z * [new branch] gh/xuanzhang816/34/base -> origin/gh/xuanzhang816/34/base 2025-12-04T09:16:47.8682856Z * [new branch] gh/xuanzhang816/34/head -> origin/gh/xuanzhang816/34/head 2025-12-04T09:16:47.8684663Z * [new branch] gh/xuanzhang816/34/orig -> origin/gh/xuanzhang816/34/orig 2025-12-04T09:16:47.8687735Z * [new branch] gh/xuanzhang816/35/base -> origin/gh/xuanzhang816/35/base 2025-12-04T09:16:47.8689282Z * [new branch] gh/xuanzhang816/35/head -> origin/gh/xuanzhang816/35/head 2025-12-04T09:16:47.8691268Z * [new branch] gh/xuanzhang816/35/orig -> origin/gh/xuanzhang816/35/orig 2025-12-04T09:16:47.8694626Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-12-04T09:16:47.8696228Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-12-04T09:16:47.8697994Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-12-04T09:16:47.8700609Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-12-04T09:16:47.8702479Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-12-04T09:16:47.8704250Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-12-04T09:16:47.8706795Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-12-04T09:16:47.8708617Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-12-04T09:16:47.8710581Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-12-04T09:16:47.8713017Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-12-04T09:16:47.8714871Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-12-04T09:16:47.8716687Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-12-04T09:16:47.8719128Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-12-04T09:16:47.8720962Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-12-04T09:16:47.8722754Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-12-04T09:16:47.8725299Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-12-04T09:16:47.8727661Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-12-04T09:16:47.8729528Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-12-04T09:16:47.8732021Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-12-04T09:16:47.8733847Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-12-04T09:16:47.8735864Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-12-04T09:16:47.8739740Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-12-04T09:16:47.8741594Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-12-04T09:16:47.8743411Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-12-04T09:16:47.8745901Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-12-04T09:16:47.8747771Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-12-04T09:16:47.8750373Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-12-04T09:16:47.8752726Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-12-04T09:16:47.8755139Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-12-04T09:16:47.8757660Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-12-04T09:16:47.8759534Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-12-04T09:16:47.8761389Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-12-04T09:16:47.8763940Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-12-04T09:16:47.8765782Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-12-04T09:16:47.8767824Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-12-04T09:16:47.8770366Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-12-04T09:16:47.8772196Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-12-04T09:16:47.8774011Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-12-04T09:16:47.8776537Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-12-04T09:16:47.8778490Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-12-04T09:16:47.8780300Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-12-04T09:16:47.8783453Z * [new branch] gh/yang-yu-hang/1/base -> origin/gh/yang-yu-hang/1/base 2025-12-04T09:16:47.8785428Z * [new branch] gh/yang-yu-hang/1/head -> origin/gh/yang-yu-hang/1/head 2025-12-04T09:16:47.8787420Z * [new branch] gh/yang-yu-hang/1/orig -> origin/gh/yang-yu-hang/1/orig 2025-12-04T09:16:47.8789937Z * [new branch] gh/yang-yu-hang/2/base -> origin/gh/yang-yu-hang/2/base 2025-12-04T09:16:47.8792007Z * [new branch] gh/yang-yu-hang/2/head -> origin/gh/yang-yu-hang/2/head 2025-12-04T09:16:47.8794058Z * [new branch] gh/yang-yu-hang/2/orig -> origin/gh/yang-yu-hang/2/orig 2025-12-04T09:16:47.8796622Z * [new branch] gh/yang-yu-hang/3/base -> origin/gh/yang-yu-hang/3/base 2025-12-04T09:16:47.8798557Z * [new branch] gh/yang-yu-hang/3/head -> origin/gh/yang-yu-hang/3/head 2025-12-04T09:16:47.8800489Z * [new branch] gh/yang-yu-hang/3/orig -> origin/gh/yang-yu-hang/3/orig 2025-12-04T09:16:47.8803584Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-12-04T09:16:47.8805480Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-12-04T09:16:47.8807282Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-12-04T09:16:47.8809973Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-12-04T09:16:47.8811816Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-12-04T09:16:47.8813652Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-12-04T09:16:47.8816138Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-12-04T09:16:47.8817929Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-12-04T09:16:47.8819769Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-12-04T09:16:47.8822347Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-12-04T09:16:47.8824113Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-12-04T09:16:47.8825952Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-12-04T09:16:47.8828991Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-12-04T09:16:47.8830887Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-12-04T09:16:47.8832775Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-12-04T09:16:47.8835499Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-12-04T09:16:47.8837555Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-12-04T09:16:47.8839469Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-12-04T09:16:47.8841880Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-12-04T09:16:47.8843845Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-12-04T09:16:47.8845521Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-12-04T09:16:47.8848805Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-12-04T09:16:47.8850557Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-12-04T09:16:47.8852405Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-12-04T09:16:47.8854977Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-12-04T09:16:47.8856779Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-12-04T09:16:47.8858695Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-12-04T09:16:47.8861347Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-12-04T09:16:47.8863230Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-12-04T09:16:47.8864994Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-12-04T09:16:47.8867525Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-12-04T09:16:47.8869352Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-12-04T09:16:47.8871258Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-12-04T09:16:47.8873806Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-12-04T09:16:47.8875689Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-12-04T09:16:47.8877614Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-12-04T09:16:47.8880145Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-12-04T09:16:47.8881935Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-12-04T09:16:47.8884254Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-12-04T09:16:47.8886727Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-12-04T09:16:47.8888720Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-12-04T09:16:47.8890586Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-12-04T09:16:47.8893069Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-12-04T09:16:47.8894964Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-12-04T09:16:47.8896776Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-12-04T09:16:47.8899411Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-12-04T09:16:47.8901193Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-12-04T09:16:47.8903051Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-12-04T09:16:47.8905907Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-12-04T09:16:47.8907724Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-12-04T09:16:47.8909542Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-12-04T09:16:47.8912171Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-12-04T09:16:47.8913939Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-12-04T09:16:47.8915782Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-12-04T09:16:47.8918195Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-12-04T09:16:47.8920196Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-12-04T09:16:47.8921901Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-12-04T09:16:47.8924257Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-12-04T09:16:47.8926102Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-12-04T09:16:47.8928085Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-12-04T09:16:47.8930544Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-12-04T09:16:47.8932274Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-12-04T09:16:47.8934036Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-12-04T09:16:47.8939554Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-12-04T09:16:47.8941446Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-12-04T09:16:47.8952136Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-12-04T09:16:47.8952689Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-12-04T09:16:47.8953204Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-12-04T09:16:47.8953712Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-12-04T09:16:47.8954199Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-12-04T09:16:47.8954699Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-12-04T09:16:47.8955776Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-12-04T09:16:47.8958596Z * [new branch] gh/ydwu4/339/base -> origin/gh/ydwu4/339/base 2025-12-04T09:16:47.8960422Z * [new branch] gh/ydwu4/339/head -> origin/gh/ydwu4/339/head 2025-12-04T09:16:47.8962207Z * [new branch] gh/ydwu4/339/orig -> origin/gh/ydwu4/339/orig 2025-12-04T09:16:47.8965275Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-12-04T09:16:47.8967123Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-12-04T09:16:47.8969814Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-12-04T09:16:47.8971607Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-12-04T09:16:47.8975757Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-12-04T09:16:47.8977989Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-12-04T09:16:47.8979940Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-12-04T09:16:47.8982428Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-12-04T09:16:47.8984385Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-12-04T09:16:47.8986217Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-12-04T09:16:47.8989868Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-12-04T09:16:47.8991696Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-12-04T09:16:47.8994101Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-12-04T09:16:47.8995863Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-12-04T09:16:47.8998981Z * [new branch] gh/yushangdi/1/base -> origin/gh/yushangdi/1/base 2025-12-04T09:16:47.9000989Z * [new branch] gh/yushangdi/1/head -> origin/gh/yushangdi/1/head 2025-12-04T09:16:47.9003458Z * [new branch] gh/yushangdi/10/base -> origin/gh/yushangdi/10/base 2025-12-04T09:16:47.9005282Z * [new branch] gh/yushangdi/10/head -> origin/gh/yushangdi/10/head 2025-12-04T09:16:47.9007077Z * [new branch] gh/yushangdi/10/orig -> origin/gh/yushangdi/10/orig 2025-12-04T09:16:47.9009796Z * [new branch] gh/yushangdi/11/base -> origin/gh/yushangdi/11/base 2025-12-04T09:16:47.9011601Z * [new branch] gh/yushangdi/11/head -> origin/gh/yushangdi/11/head 2025-12-04T09:16:47.9013416Z * [new branch] gh/yushangdi/11/orig -> origin/gh/yushangdi/11/orig 2025-12-04T09:16:47.9015774Z * [new branch] gh/yushangdi/2/base -> origin/gh/yushangdi/2/base 2025-12-04T09:16:47.9017521Z * [new branch] gh/yushangdi/2/head -> origin/gh/yushangdi/2/head 2025-12-04T09:16:47.9020127Z * [new branch] gh/yushangdi/7/base -> origin/gh/yushangdi/7/base 2025-12-04T09:16:47.9021904Z * [new branch] gh/yushangdi/7/head -> origin/gh/yushangdi/7/head 2025-12-04T09:16:47.9023731Z * [new branch] gh/yushangdi/7/orig -> origin/gh/yushangdi/7/orig 2025-12-04T09:16:47.9027078Z * [new branch] gh/yushangdi/8/base -> origin/gh/yushangdi/8/base 2025-12-04T09:16:47.9029047Z * [new branch] gh/yushangdi/8/head -> origin/gh/yushangdi/8/head 2025-12-04T09:16:47.9030922Z * [new branch] gh/yushangdi/8/orig -> origin/gh/yushangdi/8/orig 2025-12-04T09:16:47.9033279Z * [new branch] gh/yushangdi/9/base -> origin/gh/yushangdi/9/base 2025-12-04T09:16:47.9035098Z * [new branch] gh/yushangdi/9/head -> origin/gh/yushangdi/9/head 2025-12-04T09:16:47.9037416Z * [new branch] gh/yushangdi/9/orig -> origin/gh/yushangdi/9/orig 2025-12-04T09:16:47.9040542Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-12-04T09:16:47.9042340Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-12-04T09:16:47.9044141Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-12-04T09:16:47.9046767Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-12-04T09:16:47.9048761Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-12-04T09:16:47.9050582Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-12-04T09:16:47.9053201Z * [new branch] gh/zklaus/21/base -> origin/gh/zklaus/21/base 2025-12-04T09:16:47.9055034Z * [new branch] gh/zklaus/21/head -> origin/gh/zklaus/21/head 2025-12-04T09:16:47.9056854Z * [new branch] gh/zklaus/21/orig -> origin/gh/zklaus/21/orig 2025-12-04T09:16:47.9059284Z * [new branch] gh/zklaus/22/base -> origin/gh/zklaus/22/base 2025-12-04T09:16:47.9061088Z * [new branch] gh/zklaus/22/head -> origin/gh/zklaus/22/head 2025-12-04T09:16:47.9063391Z * [new branch] gh/zklaus/22/orig -> origin/gh/zklaus/22/orig 2025-12-04T09:16:47.9065901Z * [new branch] gh/zklaus/23/base -> origin/gh/zklaus/23/base 2025-12-04T09:16:47.9067762Z * [new branch] gh/zklaus/23/head -> origin/gh/zklaus/23/head 2025-12-04T09:16:47.9069569Z * [new branch] gh/zklaus/23/orig -> origin/gh/zklaus/23/orig 2025-12-04T09:16:47.9072096Z * [new branch] gh/zklaus/24/base -> origin/gh/zklaus/24/base 2025-12-04T09:16:47.9074577Z * [new branch] gh/zklaus/24/head -> origin/gh/zklaus/24/head 2025-12-04T09:16:47.9076421Z * [new branch] gh/zklaus/24/orig -> origin/gh/zklaus/24/orig 2025-12-04T09:16:47.9079829Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-12-04T09:16:47.9081491Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-12-04T09:16:47.9083274Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-12-04T09:16:47.9086446Z * [new branch] gh/zou3519/1199/base -> origin/gh/zou3519/1199/base 2025-12-04T09:16:47.9088402Z * [new branch] gh/zou3519/1199/head -> origin/gh/zou3519/1199/head 2025-12-04T09:16:47.9090355Z * [new branch] gh/zou3519/1199/orig -> origin/gh/zou3519/1199/orig 2025-12-04T09:16:47.9092773Z * [new branch] gh/zou3519/1200/base -> origin/gh/zou3519/1200/base 2025-12-04T09:16:47.9094585Z * [new branch] gh/zou3519/1200/head -> origin/gh/zou3519/1200/head 2025-12-04T09:16:47.9096417Z * [new branch] gh/zou3519/1200/orig -> origin/gh/zou3519/1200/orig 2025-12-04T09:16:47.9099052Z * [new branch] gh/zou3519/1201/base -> origin/gh/zou3519/1201/base 2025-12-04T09:16:47.9100749Z * [new branch] gh/zou3519/1201/head -> origin/gh/zou3519/1201/head 2025-12-04T09:16:47.9102676Z * [new branch] gh/zou3519/1201/orig -> origin/gh/zou3519/1201/orig 2025-12-04T09:16:47.9105024Z * [new branch] gh/zou3519/1202/base -> origin/gh/zou3519/1202/base 2025-12-04T09:16:47.9106969Z * [new branch] gh/zou3519/1202/head -> origin/gh/zou3519/1202/head 2025-12-04T09:16:47.9108789Z * [new branch] gh/zou3519/1202/orig -> origin/gh/zou3519/1202/orig 2025-12-04T09:16:47.9111931Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-12-04T09:16:47.9113787Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-12-04T09:16:47.9116402Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-12-04T09:16:47.9118199Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-12-04T09:16:47.9120070Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-12-04T09:16:47.9122984Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-12-04T09:16:47.9125327Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-12-04T09:16:47.9127222Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-12-04T09:16:47.9129969Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-12-04T09:16:47.9131707Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-12-04T09:16:47.9133515Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-12-04T09:16:47.9136371Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-12-04T09:16:47.9138117Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-12-04T09:16:47.9139987Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-12-04T09:16:47.9142700Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-12-04T09:16:47.9144475Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-12-04T09:16:47.9146314Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-12-04T09:16:47.9148822Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-12-04T09:16:47.9150700Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-12-04T09:16:47.9153766Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-12-04T09:16:47.9155874Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-12-04T09:16:47.9157563Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-12-04T09:16:47.9160337Z * [new branch] gh/zpcore/22/base -> origin/gh/zpcore/22/base 2025-12-04T09:16:47.9162105Z * [new branch] gh/zpcore/22/head -> origin/gh/zpcore/22/head 2025-12-04T09:16:47.9163918Z * [new branch] gh/zpcore/22/orig -> origin/gh/zpcore/22/orig 2025-12-04T09:16:47.9166482Z * [new branch] gh/zpcore/23/base -> origin/gh/zpcore/23/base 2025-12-04T09:16:47.9168480Z * [new branch] gh/zpcore/23/head -> origin/gh/zpcore/23/head 2025-12-04T09:16:47.9170373Z * [new branch] gh/zpcore/23/orig -> origin/gh/zpcore/23/orig 2025-12-04T09:16:47.9172738Z * [new branch] gh/zpcore/24/base -> origin/gh/zpcore/24/base 2025-12-04T09:16:47.9174560Z * [new branch] gh/zpcore/24/head -> origin/gh/zpcore/24/head 2025-12-04T09:16:47.9176385Z * [new branch] gh/zpcore/24/orig -> origin/gh/zpcore/24/orig 2025-12-04T09:16:47.9179264Z * [new branch] gh/zpcore/25/base -> origin/gh/zpcore/25/base 2025-12-04T09:16:47.9181095Z * [new branch] gh/zpcore/25/head -> origin/gh/zpcore/25/head 2025-12-04T09:16:47.9182877Z * [new branch] gh/zpcore/25/orig -> origin/gh/zpcore/25/orig 2025-12-04T09:16:47.9185469Z * [new branch] gh/zpcore/26/base -> origin/gh/zpcore/26/base 2025-12-04T09:16:47.9187440Z * [new branch] gh/zpcore/26/head -> origin/gh/zpcore/26/head 2025-12-04T09:16:47.9189735Z * [new branch] gh/zpcore/26/orig -> origin/gh/zpcore/26/orig 2025-12-04T09:16:47.9192377Z * [new branch] gh/zpcore/27/base -> origin/gh/zpcore/27/base 2025-12-04T09:16:47.9194215Z * [new branch] gh/zpcore/27/head -> origin/gh/zpcore/27/head 2025-12-04T09:16:47.9195997Z * [new branch] gh/zpcore/27/orig -> origin/gh/zpcore/27/orig 2025-12-04T09:16:47.9199044Z * [new branch] gh/zpcore/28/base -> origin/gh/zpcore/28/base 2025-12-04T09:16:47.9201336Z * [new branch] gh/zpcore/28/head -> origin/gh/zpcore/28/head 2025-12-04T09:16:47.9203122Z * [new branch] gh/zpcore/28/orig -> origin/gh/zpcore/28/orig 2025-12-04T09:16:47.9205576Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-12-04T09:16:47.9207342Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-12-04T09:16:47.9210012Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-12-04T09:16:47.9212180Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-12-04T09:16:47.9214545Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-12-04T09:16:47.9216314Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-12-04T09:16:47.9218766Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-12-04T09:16:47.9220548Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-12-04T09:16:47.9223421Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-12-04T09:16:47.9225166Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-12-04T09:16:47.9227711Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-12-04T09:16:47.9229535Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-12-04T09:16:47.9231582Z * [new branch] google-main -> origin/google-main 2025-12-04T09:16:47.9234202Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-12-04T09:16:47.9235991Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-12-04T09:16:47.9238969Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-12-04T09:16:47.9241449Z * [new branch] hameerabbasi/complex_tensor_subclass -> origin/hameerabbasi/complex_tensor_subclass 2025-12-04T09:16:47.9243353Z * [new branch] hameerabbasi/fix-ctensor-gradcheck-tests -> origin/hameerabbasi/fix-ctensor-gradcheck-tests 2025-12-04T09:16:47.9245092Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-12-04T09:16:47.9246840Z * [new branch] hc_baseline -> origin/hc_baseline 2025-12-04T09:16:47.9248984Z * [new branch] hhh_rand -> origin/hhh_rand 2025-12-04T09:16:47.9251430Z * [new branch] huba/f1 -> origin/huba/f1 2025-12-04T09:16:47.9254012Z * [new branch] increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test -> origin/increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test 2025-12-04T09:16:47.9255400Z * [new branch] inlining -> origin/inlining 2025-12-04T09:16:47.9257456Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-12-04T09:16:47.9259423Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-12-04T09:16:47.9261644Z * [new branch] instrument-trunk-pull-linux-with-job-test-filters -> origin/instrument-trunk-pull-linux-with-job-test-filters 2025-12-04T09:16:47.9263052Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-12-04T09:16:47.9265196Z * [new branch] issue#58739 -> origin/issue#58739 2025-12-04T09:16:47.9267188Z * [new branch] jainapurva-patch-1 -> origin/jainapurva-patch-1 2025-12-04T09:16:47.9269686Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-12-04T09:16:47.9271438Z * [new branch] jathu/sve -> origin/jathu/sve 2025-12-04T09:16:47.9274133Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-12-04T09:16:47.9275892Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-12-04T09:16:47.9278352Z * [new branch] jiannanWang/memorysnapshot_filter -> origin/jiannanWang/memorysnapshot_filter 2025-12-04T09:16:47.9280323Z * [new branch] jiannanWang/profilerstepwarning -> origin/jiannanWang/profilerstepwarning 2025-12-04T09:16:47.9282180Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-12-04T09:16:47.9284200Z * [new branch] jithunnair-amd-patch-10 -> origin/jithunnair-amd-patch-10 2025-12-04T09:16:47.9286640Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-12-04T09:16:47.9288753Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-12-04T09:16:47.9290733Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-12-04T09:16:47.9292631Z * [new branch] jithunnair-amd-patch-5 -> origin/jithunnair-amd-patch-5 2025-12-04T09:16:47.9294551Z * [new branch] jithunnair-amd-patch-6 -> origin/jithunnair-amd-patch-6 2025-12-04T09:16:47.9296441Z * [new branch] jithunnair-amd-patch-7 -> origin/jithunnair-amd-patch-7 2025-12-04T09:16:47.9298421Z * [new branch] jithunnair-amd-patch-8 -> origin/jithunnair-amd-patch-8 2025-12-04T09:16:47.9300386Z * [new branch] jithunnair-amd-patch-9 -> origin/jithunnair-amd-patch-9 2025-12-04T09:16:47.9302901Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-12-04T09:16:47.9305515Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-12-04T09:16:47.9307250Z * [new branch] kainan_test -> origin/kainan_test 2025-12-04T09:16:47.9309233Z * [new branch] larryliu0820-patch-1 -> origin/larryliu0820-patch-1 2025-12-04T09:16:47.9311882Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-12-04T09:16:47.9314392Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-12-04T09:16:47.9316828Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-12-04T09:16:47.9318636Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-12-04T09:16:47.9320397Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-12-04T09:16:47.9322169Z * [new branch] llama4-stable -> origin/llama4-stable 2025-12-04T09:16:47.9325311Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-12-04T09:16:47.9328425Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-12-04T09:16:47.9330181Z * [new branch] lucaskabela/fix_164876 -> origin/lucaskabela/fix_164876 2025-12-04T09:16:47.9332029Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-12-04T09:16:47.9333821Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-12-04T09:16:47.9335798Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-12-04T09:16:47.9340079Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-12-04T09:16:47.9342071Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-12-04T09:16:47.9344440Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-12-04T09:16:47.9346184Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-12-04T09:16:47.9348056Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-12-04T09:16:47.9349920Z * [new branch] lucaskabela/typing_ctx_manager -> origin/lucaskabela/typing_ctx_manager 2025-12-04T09:16:47.9351776Z * [new branch] lucaskabela/typing_nn_module -> origin/lucaskabela/typing_nn_module 2025-12-04T09:16:47.9353570Z * [new branch] lucaskabela/typing_user_defined -> origin/lucaskabela/typing_user_defined 2025-12-04T09:16:47.9355377Z * [new branch] lucaskabela/typing_variables -> origin/lucaskabela/typing_variables 2025-12-04T09:16:47.9357136Z * [new branch] lucaskabela/typing_variables_dicts -> origin/lucaskabela/typing_variables_dicts 2025-12-04T09:16:47.9359491Z * [new branch] lucaskabela/typing_variables_functions -> origin/lucaskabela/typing_variables_functions 2025-12-04T09:16:47.9361295Z * [new branch] lucaskabela/typing_variables_lists -> origin/lucaskabela/typing_variables_lists 2025-12-04T09:16:47.9363855Z * [new branch] lw/torch_box_by_ref -> origin/lw/torch_box_by_ref 2025-12-04T09:16:47.9365762Z * [new branch] main -> origin/main 2025-12-04T09:16:47.9368005Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-12-04T09:16:47.9370049Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-12-04T09:16:47.9371990Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-12-04T09:16:47.9374171Z * [new branch] malfet-patch-4 -> origin/malfet-patch-4 2025-12-04T09:16:47.9376099Z * [new branch] malfet-patch-5 -> origin/malfet-patch-5 2025-12-04T09:16:47.9377960Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-12-04T09:16:47.9379927Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-12-04T09:16:47.9381982Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-12-04T09:16:47.9384600Z * [new branch] malfet/add-3.14-ci -> origin/malfet/add-3.14-ci 2025-12-04T09:16:47.9386539Z * [new branch] malfet/be-do-not-make-typos-in-build-artifacts -> origin/malfet/be-do-not-make-typos-in-build-artifacts 2025-12-04T09:16:47.9388199Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-12-04T09:16:47.9390368Z * [new branch] malfet/be-remove-misisng-neon-headers -> origin/malfet/be-remove-misisng-neon-headers 2025-12-04T09:16:47.9392366Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-12-04T09:16:47.9395064Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-12-04T09:16:47.9396739Z * [new branch] manuel/inductor_link_openmp -> origin/manuel/inductor_link_openmp 2025-12-04T09:16:47.9399188Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-12-04T09:16:47.9401187Z * [new branch] mem_profiler_flaky_fix -> origin/mem_profiler_flaky_fix 2025-12-04T09:16:47.9403185Z * [new branch] mem_profiler_stack_trace -> origin/mem_profiler_stack_trace 2025-12-04T09:16:47.9405206Z * [new branch] memory_profiler_stack -> origin/memory_profiler_stack 2025-12-04T09:16:47.9407210Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-12-04T09:16:47.9409287Z * [new branch] mingw_posix -> origin/mingw_posix 2025-12-04T09:16:47.9411854Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-12-04T09:16:47.9413873Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-12-04T09:16:47.9415643Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-12-04T09:16:47.9417391Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-12-04T09:16:47.9419142Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-12-04T09:16:47.9420972Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-12-04T09:16:47.9422733Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-12-04T09:16:47.9424436Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-12-04T09:16:47.9426013Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-12-04T09:16:47.9428120Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-12-04T09:16:47.9430195Z * [new branch] mlazos/bwd -> origin/mlazos/bwd 2025-12-04T09:16:47.9431978Z * [new branch] mlazos/combo-test -> origin/mlazos/combo-test 2025-12-04T09:16:47.9433817Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-12-04T09:16:47.9435964Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-12-04T09:16:47.9438153Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-12-04T09:16:47.9440044Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-12-04T09:16:47.9441941Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-12-04T09:16:47.9443960Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-12-04T09:16:47.9445723Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-12-04T09:16:47.9447546Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-12-04T09:16:47.9449516Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-12-04T09:16:47.9451388Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-12-04T09:16:47.9453275Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-12-04T09:16:47.9455108Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-12-04T09:16:47.9456984Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-12-04T09:16:47.9458961Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-12-04T09:16:47.9460922Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-12-04T09:16:47.9462605Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-12-04T09:16:47.9464440Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-12-04T09:16:47.9466317Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-12-04T09:16:47.9468240Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-12-04T09:16:47.9470055Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-12-04T09:16:47.9471916Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-12-04T09:16:47.9473773Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-12-04T09:16:47.9475723Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-12-04T09:16:47.9477531Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-12-04T09:16:47.9479405Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-12-04T09:16:47.9481886Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-12-04T09:16:47.9483744Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-12-04T09:16:47.9485596Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-12-04T09:16:47.9487483Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-12-04T09:16:47.9489545Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-12-04T09:16:47.9491381Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-12-04T09:16:47.9493196Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-12-04T09:16:47.9495073Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-12-04T09:16:47.9496899Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-12-04T09:16:47.9498675Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-12-04T09:16:47.9500589Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-12-04T09:16:47.9502397Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-12-04T09:16:47.9504242Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-12-04T09:16:47.9506067Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-12-04T09:16:47.9507886Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-12-04T09:16:47.9509704Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-12-04T09:16:47.9511754Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-12-04T09:16:47.9513507Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-12-04T09:16:47.9515325Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-12-04T09:16:47.9517205Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-12-04T09:16:47.9519013Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-12-04T09:16:47.9521195Z * [new branch] mlazos/main -> origin/mlazos/main 2025-12-04T09:16:47.9523051Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-12-04T09:16:47.9524982Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-12-04T09:16:47.9527611Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-12-04T09:16:47.9529619Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-12-04T09:16:47.9531397Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-12-04T09:16:47.9533313Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-12-04T09:16:47.9535154Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-12-04T09:16:47.9537145Z * [new branch] mlazos/overguarding -> origin/mlazos/overguarding 2025-12-04T09:16:47.9539021Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-12-04T09:16:47.9540919Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-12-04T09:16:47.9542756Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-12-04T09:16:47.9544627Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-12-04T09:16:47.9546491Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-12-04T09:16:47.9548336Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-12-04T09:16:47.9550243Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-12-04T09:16:47.9552159Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-12-04T09:16:47.9554029Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-12-04T09:16:47.9555669Z * [new branch] mlazos/stests -> origin/mlazos/stests 2025-12-04T09:16:47.9557556Z * [new branch] mlazos/stream-ops -> origin/mlazos/stream-ops 2025-12-04T09:16:47.9559376Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-12-04T09:16:47.9561277Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-12-04T09:16:47.9563076Z * [new branch] mlazos/test -> origin/mlazos/test 2025-12-04T09:16:47.9564994Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-12-04T09:16:47.9566875Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-12-04T09:16:47.9568908Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-12-04T09:16:47.9571015Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-12-04T09:16:47.9572924Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-12-04T09:16:47.9574689Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-12-04T09:16:47.9576580Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-12-04T09:16:47.9578409Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-12-04T09:16:47.9580467Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-12-04T09:16:47.9582233Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-12-04T09:16:47.9584114Z * [new branch] mlazos/user-stream-base -> origin/mlazos/user-stream-base 2025-12-04T09:16:47.9585974Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-12-04T09:16:47.9587913Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-12-04T09:16:47.9589720Z * [new branch] mlazos/user-streams-backup2 -> origin/mlazos/user-streams-backup2 2025-12-04T09:16:47.9591570Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-12-04T09:16:47.9593466Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-12-04T09:16:47.9595320Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-12-04T09:16:47.9597231Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-12-04T09:16:47.9599160Z * [new branch] module-shim -> origin/module-shim 2025-12-04T09:16:47.9601214Z * [new branch] move_config -> origin/move_config 2025-12-04T09:16:47.9603797Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-12-04T09:16:47.9606269Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-12-04T09:16:47.9609049Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-12-04T09:16:47.9610894Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-12-04T09:16:47.9612803Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-12-04T09:16:47.9614681Z * [new branch] new-codegen -> origin/new-codegen 2025-12-04T09:16:47.9616649Z * [new branch] newtest-base -> origin/newtest-base 2025-12-04T09:16:47.9619420Z * [new branch] ngimel/addmm_dtype -> origin/ngimel/addmm_dtype 2025-12-04T09:16:47.9621305Z * [new branch] ngimel/div_inv -> origin/ngimel/div_inv 2025-12-04T09:16:47.9623446Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-12-04T09:16:47.9625685Z * [new branch] ngimel/gather_grid -> origin/ngimel/gather_grid 2025-12-04T09:16:47.9627673Z * [new branch] ngimel/gather_grid_release -> origin/ngimel/gather_grid_release 2025-12-04T09:16:47.9629359Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-12-04T09:16:47.9631184Z * [new branch] ngimel/hostalloc -> origin/ngimel/hostalloc 2025-12-04T09:16:47.9632903Z * [new branch] ngimel/storage_id -> origin/ngimel/storage_id 2025-12-04T09:16:47.9634872Z * [new branch] nightly -> origin/nightly 2025-12-04T09:16:47.9638018Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-12-04T09:16:47.9639776Z * [new branch] nikitaved/addmm_epilogue_fusions_2d_bias -> origin/nikitaved/addmm_epilogue_fusions_2d_bias 2025-12-04T09:16:47.9641511Z * [new branch] nikitaved/addmm_epilogue_fusions_inductor -> origin/nikitaved/addmm_epilogue_fusions_inductor 2025-12-04T09:16:47.9643570Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-12-04T09:16:47.9645685Z * [new branch] nikitaved/grad_addmm_epilogue_fusions -> origin/nikitaved/grad_addmm_epilogue_fusions 2025-12-04T09:16:47.9647919Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-12-04T09:16:47.9649989Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-12-04T09:16:47.9652145Z * [new branch] nmacchioni-perf-test-async-autotune -> origin/nmacchioni-perf-test-async-autotune 2025-12-04T09:16:47.9653968Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-12-04T09:16:47.9655902Z * [new branch] nofun-hack -> origin/nofun-hack 2025-12-04T09:16:47.9657942Z * [new branch] norm_bench -> origin/norm_bench 2025-12-04T09:16:47.9660521Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-12-04T09:16:47.9662377Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-12-04T09:16:47.9664798Z * [new branch] optimizer_test -> origin/optimizer_test 2025-12-04T09:16:47.9668091Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-12-04T09:16:47.9670050Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-12-04T09:16:47.9671861Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-12-04T09:16:47.9673923Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-12-04T09:16:47.9675801Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-12-04T09:16:47.9677871Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-12-04T09:16:47.9679760Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-12-04T09:16:47.9681675Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-12-04T09:16:47.9683539Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-12-04T09:16:47.9685523Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-12-04T09:16:47.9687326Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-12-04T09:16:47.9689347Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-12-04T09:16:47.9691229Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-12-04T09:16:47.9693035Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-12-04T09:16:47.9694917Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-12-04T09:16:47.9697066Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-12-04T09:16:47.9699453Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-12-04T09:16:47.9701272Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-12-04T09:16:47.9705498Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-12-04T09:16:47.9707250Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-12-04T09:16:47.9711011Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-12-04T09:16:47.9712995Z * [new branch] oulgen-patch-1 -> origin/oulgen-patch-1 2025-12-04T09:16:47.9715020Z * [new branch] oulgen-patch-2 -> origin/oulgen-patch-2 2025-12-04T09:16:47.9717297Z * [new branch] oulgen-patch-3 -> origin/oulgen-patch-3 2025-12-04T09:16:47.9719343Z * [new branch] oulgen-patch-4 -> origin/oulgen-patch-4 2025-12-04T09:16:47.9721463Z * [new branch] padded-tensor -> origin/padded-tensor 2025-12-04T09:16:47.9723493Z * [new branch] pca2 -> origin/pca2 2025-12-04T09:16:47.9725580Z * [new branch] per_channel_backup -> origin/per_channel_backup 2025-12-04T09:16:47.9727886Z * [new branch] perf_ops -> origin/perf_ops 2025-12-04T09:16:47.9729694Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-12-04T09:16:47.9731880Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-12-04T09:16:47.9734453Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-12-04T09:16:47.9736417Z * [new branch] pianpwk/_debug_mode_for_triton_draft -> origin/pianpwk/_debug_mode_for_triton_draft 2025-12-04T09:16:47.9740072Z * [new branch] pianpwk/_debug_nn_module_compile -> origin/pianpwk/_debug_nn_module_compile 2025-12-04T09:16:47.9741874Z * [new branch] pianpwk/_draft_triton_11_3 -> origin/pianpwk/_draft_triton_11_3 2025-12-04T09:16:47.9743592Z * [new branch] pianpwk/_manual_bucket_draft -> origin/pianpwk/_manual_bucket_draft 2025-12-04T09:16:47.9745787Z * [new branch] pianpwk/_profile_w_dispatch_keys -> origin/pianpwk/_profile_w_dispatch_keys 2025-12-04T09:16:47.9747941Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-12-04T09:16:47.9749926Z * [new branch] pianpwk/_unbacked_local_shard_size -> origin/pianpwk/_unbacked_local_shard_size 2025-12-04T09:16:47.9751715Z * [new branch] pianpwk/anomaly_tb -> origin/pianpwk/anomaly_tb 2025-12-04T09:16:47.9753595Z * [new branch] pianpwk/auto_fx_annotate -> origin/pianpwk/auto_fx_annotate 2025-12-04T09:16:47.9755592Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-12-04T09:16:47.9757822Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-12-04T09:16:47.9759785Z * [new branch] pianpwk/debug_fwd_stack_traces -> origin/pianpwk/debug_fwd_stack_traces 2025-12-04T09:16:47.9761737Z * [new branch] pianpwk/debug_hash_tensor -> origin/pianpwk/debug_hash_tensor 2025-12-04T09:16:47.9763646Z * [new branch] pianpwk/debug_mode_annotate -> origin/pianpwk/debug_mode_annotate 2025-12-04T09:16:47.9765401Z * [new branch] pianpwk/debug_mode_defaults -> origin/pianpwk/debug_mode_defaults 2025-12-04T09:16:47.9767213Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-12-04T09:16:47.9769513Z * [new branch] pianpwk/debug_mode_opcall_refactor -> origin/pianpwk/debug_mode_opcall_refactor 2025-12-04T09:16:47.9771154Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-12-04T09:16:47.9773101Z * [new branch] pianpwk/debug_mode_triton -> origin/pianpwk/debug_mode_triton 2025-12-04T09:16:47.9775579Z * [new branch] pianpwk/debug_show_stack_trace -> origin/pianpwk/debug_show_stack_trace 2025-12-04T09:16:47.9777440Z * [new branch] pianpwk/debug_wait_on_collective -> origin/pianpwk/debug_wait_on_collective 2025-12-04T09:16:47.9779340Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-12-04T09:16:47.9790478Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-12-04T09:16:47.9790912Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-12-04T09:16:47.9791187Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-12-04T09:16:47.9791501Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-12-04T09:16:47.9791765Z * [new branch] pianpwk/dtensor_custom_chunk -> origin/pianpwk/dtensor_custom_chunk 2025-12-04T09:16:47.9792048Z * [new branch] pianpwk/dtensor_unbacked_keypath -> origin/pianpwk/dtensor_unbacked_keypath 2025-12-04T09:16:47.9792925Z * [new branch] pianpwk/event_list_tree -> origin/pianpwk/event_list_tree 2025-12-04T09:16:47.9794720Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-12-04T09:16:47.9796500Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-12-04T09:16:47.9798409Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-12-04T09:16:47.9800368Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-12-04T09:16:47.9802187Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-12-04T09:16:47.9804055Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-12-04T09:16:47.9806031Z * [new branch] pianpwk/skip_python_keys_alternate -> origin/pianpwk/skip_python_keys_alternate 2025-12-04T09:16:47.9808068Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-12-04T09:16:47.9809835Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-12-04T09:16:47.9811694Z * [new branch] pianpwk/symint_one_hot -> origin/pianpwk/symint_one_hot 2025-12-04T09:16:47.9813729Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-12-04T09:16:47.9816627Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-12-04T09:16:47.9818071Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-12-04T09:16:47.9819446Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-12-04T09:16:47.9821387Z * [new branch] pianpwk/unbacked_dtensor_mm -> origin/pianpwk/unbacked_dtensor_mm 2025-12-04T09:16:47.9823389Z * [new branch] pianpwk/unbacked_tracing_12_2 -> origin/pianpwk/unbacked_tracing_12_2 2025-12-04T09:16:47.9825228Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-12-04T09:16:47.9827084Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-12-04T09:16:47.9829755Z * [new branch] piz/fix_partial_backward_1112 -> origin/piz/fix_partial_backward_1112 2025-12-04T09:16:47.9831434Z * [new branch] piz/prop_cache_clean -> origin/piz/prop_cache_clean 2025-12-04T09:16:47.9833543Z * [new branch] pool-separate -> origin/pool-separate 2025-12-04T09:16:47.9835844Z * [new branch] pr-156087 -> origin/pr-156087 2025-12-04T09:16:47.9838649Z * [new branch] pr/131860 -> origin/pr/131860 2025-12-04T09:16:47.9840616Z * [new branch] predispatch_to -> origin/predispatch_to 2025-12-04T09:16:47.9842570Z * [new branch] protect-c17 -> origin/protect-c17 2025-12-04T09:16:47.9844582Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-12-04T09:16:47.9847183Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-12-04T09:16:47.9850274Z * [new branch] q1l1/fix_device_moved_constant_type_unknown -> origin/q1l1/fix_device_moved_constant_type_unknown 2025-12-04T09:16:47.9852015Z * [new branch] q1l1/fix_wrong_default_type_for_kernel_call_args -> origin/q1l1/fix_wrong_default_type_for_kernel_call_args 2025-12-04T09:16:47.9854851Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-12-04T09:16:47.9856865Z * [new branch] quote-pytest_cache -> origin/quote-pytest_cache 2025-12-04T09:16:47.9859146Z * [new branch] reland-accgrad-stream-warn -> origin/reland-accgrad-stream-warn 2025-12-04T09:16:47.9862039Z * [new branch] release/1.10 -> origin/release/1.10 2025-12-04T09:16:47.9863784Z * [new branch] release/1.11 -> origin/release/1.11 2025-12-04T09:16:47.9865590Z * [new branch] release/1.12 -> origin/release/1.12 2025-12-04T09:16:47.9867523Z * [new branch] release/1.13 -> origin/release/1.13 2025-12-04T09:16:47.9869318Z * [new branch] release/1.4 -> origin/release/1.4 2025-12-04T09:16:47.9870923Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-12-04T09:16:47.9872771Z * [new branch] release/1.5 -> origin/release/1.5 2025-12-04T09:16:47.9874688Z * [new branch] release/1.6 -> origin/release/1.6 2025-12-04T09:16:47.9876523Z * [new branch] release/1.7 -> origin/release/1.7 2025-12-04T09:16:47.9878494Z * [new branch] release/1.8 -> origin/release/1.8 2025-12-04T09:16:47.9880224Z * [new branch] release/1.9 -> origin/release/1.9 2025-12-04T09:16:47.9882087Z * [new branch] release/2.0 -> origin/release/2.0 2025-12-04T09:16:47.9884033Z * [new branch] release/2.1 -> origin/release/2.1 2025-12-04T09:16:47.9885913Z * [new branch] release/2.2 -> origin/release/2.2 2025-12-04T09:16:47.9888284Z * [new branch] release/2.3 -> origin/release/2.3 2025-12-04T09:16:47.9890671Z * [new branch] release/2.4 -> origin/release/2.4 2025-12-04T09:16:47.9893172Z * [new branch] release/2.5 -> origin/release/2.5 2025-12-04T09:16:47.9895171Z * [new branch] release/2.6 -> origin/release/2.6 2025-12-04T09:16:47.9897286Z * [new branch] release/2.7 -> origin/release/2.7 2025-12-04T09:16:47.9899219Z * [new branch] release/2.8 -> origin/release/2.8 2025-12-04T09:16:47.9901255Z * [new branch] release/2.9 -> origin/release/2.9 2025-12-04T09:16:47.9903261Z * [new branch] release_notes -> origin/release_notes 2025-12-04T09:16:47.9905260Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-12-04T09:16:47.9907424Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-12-04T09:16:47.9909299Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-12-04T09:16:47.9911071Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-12-04T09:16:47.9913525Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-12-04T09:16:47.9917338Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-12-04T09:16:47.9921153Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-12-04T09:16:47.9924892Z * [new branch] revert-152361-gh/fadara01/1/head -> origin/revert-152361-gh/fadara01/1/head 2025-12-04T09:16:47.9928756Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-12-04T09:16:47.9931094Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-12-04T09:16:47.9932937Z * [new branch] revert-hoo-invoke-subgraph -> origin/revert-hoo-invoke-subgraph 2025-12-04T09:16:47.9934951Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-12-04T09:16:47.9937140Z * [new branch] rms_norm_patch -> origin/rms_norm_patch 2025-12-04T09:16:47.9939883Z * [new branch] ruisi/fix_all_to_all_estimation -> origin/ruisi/fix_all_to_all_estimation 2025-12-04T09:16:47.9941515Z * [new branch] ruisi/fix_comm_estimation -> origin/ruisi/fix_comm_estimation 2025-12-04T09:16:47.9943346Z * [new branch] ruisi/fix_dynamic_shape_estimation -> origin/ruisi/fix_dynamic_shape_estimation 2025-12-04T09:16:47.9945047Z * [new branch] ruisi/fix_llama3_autobucketing -> origin/ruisi/fix_llama3_autobucketing 2025-12-04T09:16:47.9947196Z * [new branch] ruisi/fix_manual_bucketing_ep_pass -> origin/ruisi/fix_manual_bucketing_ep_pass 2025-12-04T09:16:47.9949413Z * [new branch] ruisi/manual_bucket_pass -> origin/ruisi/manual_bucket_pass 2025-12-04T09:16:47.9952242Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-12-04T09:16:47.9953851Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-12-04T09:16:47.9956406Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-12-04T09:16:47.9958129Z * [new branch] rzou/njt -> origin/rzou/njt 2025-12-04T09:16:47.9960002Z * [new branch] rzou/pca -> origin/rzou/pca 2025-12-04T09:16:47.9961712Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-12-04T09:16:47.9963754Z * [new branch] samplevllm -> origin/samplevllm 2025-12-04T09:16:47.9966787Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-12-04T09:16:47.9969000Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-12-04T09:16:47.9971185Z * [new branch] sapling-pr-archive-tushar00jain -> origin/sapling-pr-archive-tushar00jain 2025-12-04T09:16:47.9973031Z * [new branch] save -> origin/save 2025-12-04T09:16:47.9975034Z * [new branch] scaled_mm -> origin/scaled_mm 2025-12-04T09:16:47.9976984Z * [new branch] scan_attempt -> origin/scan_attempt 2025-12-04T09:16:47.9979718Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-12-04T09:16:47.9981777Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-12-04T09:16:47.9984311Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-12-04T09:16:47.9986359Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-12-04T09:16:47.9988319Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-12-04T09:16:47.9990449Z * [new branch] some_rocm_inductor_skips -> origin/some_rocm_inductor_skips 2025-12-04T09:16:47.9993088Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-12-04T09:16:47.9995307Z * [new branch] sparse-mm-bf16-support -> origin/sparse-mm-bf16-support 2025-12-04T09:16:47.9997251Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-12-04T09:16:47.9999219Z * [new branch] suo -> origin/suo 2025-12-04T09:16:48.0001315Z * [new branch] sve-poc -> origin/sve-poc 2025-12-04T09:16:48.0003410Z * [new branch] switch-bn -> origin/switch-bn 2025-12-04T09:16:48.0005382Z * [new branch] sy_annotation_in_autograd_hop -> origin/sy_annotation_in_autograd_hop 2025-12-04T09:16:48.0007306Z * [new branch] sy_aot_eager_record -> origin/sy_aot_eager_record 2025-12-04T09:16:48.0009479Z * [new branch] sy_custom_bucketing -> origin/sy_custom_bucketing 2025-12-04T09:16:48.0011558Z * [new branch] sy_debug_mode_test -> origin/sy_debug_mode_test 2025-12-04T09:16:48.0013619Z * [new branch] sy_deserialize -> origin/sy_deserialize 2025-12-04T09:16:48.0015562Z * [new branch] sy_dump_gm_code -> origin/sy_dump_gm_code 2025-12-04T09:16:48.0017496Z * [new branch] sy_exp -> origin/sy_exp 2025-12-04T09:16:48.0019561Z * [new branch] sy_export_annotation -> origin/sy_export_annotation 2025-12-04T09:16:48.0021491Z * [new branch] sy_invoke_subgraph -> origin/sy_invoke_subgraph 2025-12-04T09:16:48.0023670Z * [new branch] sy_kernel_bw_name -> origin/sy_kernel_bw_name 2025-12-04T09:16:48.0025682Z * [new branch] sy_multi_arch -> origin/sy_multi_arch 2025-12-04T09:16:48.0027731Z * [new branch] sy_nn_module_stack -> origin/sy_nn_module_stack 2025-12-04T09:16:48.0029734Z * [new branch] sy_original_dtensor -> origin/sy_original_dtensor 2025-12-04T09:16:48.0031696Z * [new branch] sy_profiler_cia -> origin/sy_profiler_cia 2025-12-04T09:16:48.0033761Z * [new branch] symm_mem_sync -> origin/symm_mem_sync 2025-12-04T09:16:48.0036005Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-12-04T09:16:48.0038824Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-12-04T09:16:48.0041184Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-12-04T09:16:48.0043101Z * [new branch] test-old -> origin/test-old 2025-12-04T09:16:48.0045691Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-12-04T09:16:48.0048607Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-12-04T09:16:48.0050491Z * [new branch] tianren/customOp_enable_max_autotune -> origin/tianren/customOp_enable_max_autotune 2025-12-04T09:16:48.0052218Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-12-04T09:16:48.0054153Z * [new branch] tianren/customop_collectiveop_benchmark -> origin/tianren/customop_collectiveop_benchmark 2025-12-04T09:16:48.0056316Z * [new branch] tianren/customop_collectiveop_benchmark_fix -> origin/tianren/customop_collectiveop_benchmark_fix 2025-12-04T09:16:48.0058661Z * [new branch] tianren/customop_dynamic_config -> origin/tianren/customop_dynamic_config 2025-12-04T09:16:48.0060506Z * [new branch] tianren/dynamic_range_input -> origin/tianren/dynamic_range_input 2025-12-04T09:16:48.0062420Z * [new branch] tianren/dynamic_range_input_fix -> origin/tianren/dynamic_range_input_fix 2025-12-04T09:16:48.0064769Z * [new branch] tianren/dynamic_range_input_merge -> origin/tianren/dynamic_range_input_merge 2025-12-04T09:16:48.0066728Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-12-04T09:16:48.0069104Z * [new branch] tianren/fx_codegen_dump -> origin/tianren/fx_codegen_dump 2025-12-04T09:16:48.0070980Z * [new branch] tianren/symmetric_memory -> origin/tianren/symmetric_memory 2025-12-04T09:16:48.0072796Z * [new branch] tianren/test -> origin/tianren/test 2025-12-04T09:16:48.0074810Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-12-04T09:16:48.0076703Z * [new branch] tmp -> origin/tmp 2025-12-04T09:16:48.0078756Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-12-04T09:16:48.0080749Z * [new branch] torchtitan_integration -> origin/torchtitan_integration 2025-12-04T09:16:48.0082839Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-12-04T09:16:48.0084643Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-12-04T09:16:48.0086620Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-12-04T09:16:48.0088796Z * [new branch] triton_kernel -> origin/triton_kernel 2025-12-04T09:16:48.0090851Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-12-04T09:16:48.0092807Z * [new branch] type_dec -> origin/type_dec 2025-12-04T09:16:48.0094860Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-12-04T09:16:48.0097535Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-12-04T09:16:48.0099367Z * [new branch] update-audio-commit-hash/19087141161-1916-1 -> origin/update-audio-commit-hash/19087141161-1916-1 2025-12-04T09:16:48.0101176Z * [new branch] update-audio-commit-hash/19250643381-1929-1 -> origin/update-audio-commit-hash/19250643381-1929-1 2025-12-04T09:16:48.0102980Z * [new branch] update-audio-commit-hash/19397724337-1935-1 -> origin/update-audio-commit-hash/19397724337-1935-1 2025-12-04T09:16:48.0104719Z * [new branch] update-audio-commit-hash/19555670148-1941-1 -> origin/update-audio-commit-hash/19555670148-1941-1 2025-12-04T09:16:48.0106922Z * [new branch] update-audio-commit-hash/19750627930-1946-1 -> origin/update-audio-commit-hash/19750627930-1946-1 2025-12-04T09:16:48.0109479Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-12-04T09:16:48.0112069Z * [new branch] update-vision-commit-hash/19087141161-1916-1 -> origin/update-vision-commit-hash/19087141161-1916-1 2025-12-04T09:16:48.0113931Z * [new branch] update-vision-commit-hash/19184897099-1925-1 -> origin/update-vision-commit-hash/19184897099-1925-1 2025-12-04T09:16:48.0115670Z * [new branch] update-vision-commit-hash/19250643381-1929-1 -> origin/update-vision-commit-hash/19250643381-1929-1 2025-12-04T09:16:48.0117570Z * [new branch] update-vision-commit-hash/19381328640-1934-1 -> origin/update-vision-commit-hash/19381328640-1934-1 2025-12-04T09:16:48.0119319Z * [new branch] update-vision-commit-hash/19485237164-1938-1 -> origin/update-vision-commit-hash/19485237164-1938-1 2025-12-04T09:16:48.0122573Z * [new branch] update-vllm-commit-hash/18451675449-1879-1 -> origin/update-vllm-commit-hash/18451675449-1879-1 2025-12-04T09:16:48.0124331Z * [new branch] update-vllm-dockerfile -> origin/update-vllm-dockerfile 2025-12-04T09:16:48.0127051Z * [new branch] update-xla-commit-hash/19224287370-211-1 -> origin/update-xla-commit-hash/19224287370-211-1 2025-12-04T09:16:48.0129070Z * [new branch] update-xla-commit-hash/19422028566-212-1 -> origin/update-xla-commit-hash/19422028566-212-1 2025-12-04T09:16:48.0130856Z * [new branch] update-xla-commit-hash/19626841311-213-1 -> origin/update-xla-commit-hash/19626841311-213-1 2025-12-04T09:16:48.0132874Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-12-04T09:16:48.0134690Z * [new branch] update_operator_readme -> origin/update_operator_readme 2025-12-04T09:16:48.0139211Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-12-04T09:16:48.0141232Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-12-04T09:16:48.0143134Z * [new branch] update_slow_tests_1762155677 -> origin/update_slow_tests_1762155677 2025-12-04T09:16:48.0145230Z * [new branch] update_slow_tests_1763365283 -> origin/update_slow_tests_1763365283 2025-12-04T09:16:48.0147091Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-12-04T09:16:48.0149105Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-12-04T09:16:48.0151081Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-12-04T09:16:48.0153008Z * [new branch] upload-tests-for-autorevert -> origin/upload-tests-for-autorevert 2025-12-04T09:16:48.0155470Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-12-04T09:16:48.0157567Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-12-04T09:16:48.0159564Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-12-04T09:16:48.0161768Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-12-04T09:16:48.0163799Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-12-04T09:16:48.0165816Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-12-04T09:16:48.0167958Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-12-04T09:16:48.0170093Z * [new branch] validate_fn -> origin/validate_fn 2025-12-04T09:16:48.0172194Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-12-04T09:16:48.0174381Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-12-04T09:16:48.0176298Z * [new branch] varlen-api -> origin/varlen-api 2025-12-04T09:16:48.0178288Z * [new branch] varlen-api-backup -> origin/varlen-api-backup 2025-12-04T09:16:48.0180208Z * [new branch] varlen_batch_invariance -> origin/varlen_batch_invariance 2025-12-04T09:16:48.0182797Z * [new branch] viable/strict -> origin/viable/strict 2025-12-04T09:16:48.0185487Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-12-04T09:16:48.0187291Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-12-04T09:16:48.0189366Z * [new branch] vllmpin -> origin/vllmpin 2025-12-04T09:16:48.0191527Z * [new branch] vscode-recommend-pyrefly -> origin/vscode-recommend-pyrefly 2025-12-04T09:16:48.0193505Z * [new branch] wdvr-patch-1 -> origin/wdvr-patch-1 2025-12-04T09:16:48.0196081Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-12-04T09:16:48.0198643Z * [new branch] whc/pei -> origin/whc/pei 2025-12-04T09:16:48.0200470Z * [new branch] whc/pp_fix -> origin/whc/pp_fix 2025-12-04T09:16:48.0202299Z * [new branch] whc/sharding -> origin/whc/sharding 2025-12-04T09:16:48.0204485Z * [new branch] whc/sharding2 -> origin/whc/sharding2 2025-12-04T09:16:48.0206227Z * [new branch] whc/uneven -> origin/whc/uneven 2025-12-04T09:16:48.0208714Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-12-04T09:16:48.0210654Z * [new branch] win_warnings -> origin/win_warnings 2025-12-04T09:16:48.0212655Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-12-04T09:16:48.0214564Z * [new branch] xmfan-war -> origin/xmfan-war 2025-12-04T09:16:48.0217124Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-12-04T09:16:48.0218939Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-12-04T09:16:48.0220920Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-12-04T09:16:48.0222387Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-12-04T09:16:48.0224158Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-12-04T09:16:48.0225873Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-12-04T09:16:48.0227630Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-12-04T09:16:48.0229686Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-12-04T09:16:48.0231954Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-12-04T09:16:48.0233883Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-12-04T09:16:48.0236048Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-12-04T09:16:48.0237969Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-12-04T09:16:48.0239837Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-12-04T09:16:48.0241731Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-12-04T09:16:48.0243572Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-12-04T09:16:48.0245487Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-12-04T09:16:48.0247396Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-12-04T09:16:48.0249393Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-12-04T09:16:48.0251200Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-12-04T09:16:48.0253043Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-12-04T09:16:48.0254942Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-12-04T09:16:48.0256814Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-12-04T09:16:48.0258821Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T09:16:48.0260638Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T09:16:48.0262346Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-12-04T09:16:48.0264152Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-12-04T09:16:48.0266069Z * [new branch] xmfan/test -> origin/xmfan/test 2025-12-04T09:16:48.0268857Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-12-04T09:16:48.0270620Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-12-04T09:16:48.0272429Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-12-04T09:16:48.0274926Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-12-04T09:16:48.0276818Z * [new branch] yiming/run_with_start_end_rng_hop -> origin/yiming/run_with_start_end_rng_hop 2025-12-04T09:16:48.0278685Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-12-04T09:16:48.0281281Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-12-04T09:16:48.0283154Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-12-04T09:16:48.0284961Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-12-04T09:16:48.0286600Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-12-04T09:16:48.0289035Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-12-04T09:16:48.0290937Z * [new branch] zb2p -> origin/zb2p 2025-12-04T09:16:48.0293010Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-12-04T09:16:48.0296625Z * [new branch] zhxchen17/ci/vllm_lora_oom -> origin/zhxchen17/ci/vllm_lora_oom 2025-12-04T09:16:48.0298441Z * [new branch] zhxchen17/ci/vllm_multimodal_oom -> origin/zhxchen17/ci/vllm_multimodal_oom 2025-12-04T09:16:48.0300177Z * [new branch] zhxchen17/ci/vllm_pin -> origin/zhxchen17/ci/vllm_pin 2025-12-04T09:16:48.0302867Z * [new branch] zhxchen17/dynamo/unsafe_drop_all_guards -> origin/zhxchen17/dynamo/unsafe_drop_all_guards 2025-12-04T09:16:48.0305784Z * [new branch] zhxchen17/export/call_override -> origin/zhxchen17/export/call_override 2025-12-04T09:16:48.0307621Z * [new branch] zhxchen17/export/codemod1 -> origin/zhxchen17/export/codemod1 2025-12-04T09:16:48.0309568Z * [new branch] zhxchen17/export/ctx_return -> origin/zhxchen17/export/ctx_return 2025-12-04T09:16:48.0311528Z * [new branch] zhxchen17/export/disable_side_effect_warn -> origin/zhxchen17/export/disable_side_effect_warn 2025-12-04T09:16:48.0313253Z * [new branch] zhxchen17/export/pytree_check -> origin/zhxchen17/export/pytree_check 2025-12-04T09:16:48.0315734Z * [new branch] zhxchen17/precompile/aoti -> origin/zhxchen17/precompile/aoti 2025-12-04T09:16:48.0317636Z * [new branch] zhxchen17/precompile/globals -> origin/zhxchen17/precompile/globals 2025-12-04T09:16:48.0319660Z * [new branch] zhxchen17/precompile/inductor_guards -> origin/zhxchen17/precompile/inductor_guards 2025-12-04T09:16:48.0322014Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-12-04T09:16:48.0323911Z * [new branch] zhxchen17/torch_export_api_update -> origin/zhxchen17/torch_export_api_update 2025-12-04T09:16:48.0326434Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-12-04T09:16:48.0329786Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-12-04T09:16:48.0332073Z * [new branch] zxiiro/c7i.2xlarge -> origin/zxiiro/c7i.2xlarge 2025-12-04T09:16:48.0333927Z * [new branch] zxiiro/c7i.2xlarge.h100 -> origin/zxiiro/c7i.2xlarge.h100 2025-12-04T09:16:48.0336025Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-12-04T09:16:48.0337929Z * [new branch] zxiiro/risc64 -> origin/zxiiro/risc64 2025-12-04T09:16:48.0339732Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-12-04T09:16:48.0341426Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-12-04T09:16:48.0342963Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-12-04T09:16:48.0344697Z * [new tag] ciflow/b200/115316 -> ciflow/b200/115316 2025-12-04T09:16:48.0346010Z * [new tag] ciflow/b200/160685 -> ciflow/b200/160685 2025-12-04T09:16:48.0347271Z * [new tag] ciflow/b200/161607 -> ciflow/b200/161607 2025-12-04T09:16:48.0348527Z * [new tag] ciflow/b200/161938 -> ciflow/b200/161938 2025-12-04T09:16:48.0349898Z * [new tag] ciflow/b200/167207 -> ciflow/b200/167207 2025-12-04T09:16:48.0351114Z * [new tag] ciflow/b200/167989 -> ciflow/b200/167989 2025-12-04T09:16:48.0352452Z * [new tag] ciflow/b200/168096 -> ciflow/b200/168096 2025-12-04T09:16:48.0353880Z * [new tag] ciflow/b200/168175 -> ciflow/b200/168175 2025-12-04T09:16:48.0355323Z * [new tag] ciflow/b200/168195 -> ciflow/b200/168195 2025-12-04T09:16:48.0356560Z * [new tag] ciflow/b200/169200 -> ciflow/b200/169200 2025-12-04T09:16:48.0357913Z * [new tag] ciflow/b200/169216 -> ciflow/b200/169216 2025-12-04T09:16:48.0359648Z * [new tag] ciflow/b200/169380 -> ciflow/b200/169380 2025-12-04T09:16:48.0361527Z * [new tag] ciflow/b200/169412 -> ciflow/b200/169412 2025-12-04T09:16:48.0363047Z * [new tag] ciflow/b200/169470 -> ciflow/b200/169470 2025-12-04T09:16:48.0364397Z * [new tag] ciflow/b200/169471 -> ciflow/b200/169471 2025-12-04T09:16:48.0365696Z * [new tag] ciflow/b200/169472 -> ciflow/b200/169472 2025-12-04T09:16:48.0367516Z * [new tag] ciflow/b200/169514 -> ciflow/b200/169514 2025-12-04T09:16:48.0368975Z * [new tag] ciflow/b200/169517 -> ciflow/b200/169517 2025-12-04T09:16:48.0370726Z * [new tag] ciflow/binaries/165922 -> ciflow/binaries/165922 2025-12-04T09:16:48.0372092Z * [new tag] ciflow/binaries/169510 -> ciflow/binaries/169510 2025-12-04T09:16:48.0373732Z * [new tag] ciflow/binaries_wheel/157994 -> ciflow/binaries_wheel/157994 2025-12-04T09:16:48.0375090Z * [new tag] ciflow/binaries_wheel/166829 -> ciflow/binaries_wheel/166829 2025-12-04T09:16:48.0376399Z * [new tag] ciflow/binaries_wheel/167972 -> ciflow/binaries_wheel/167972 2025-12-04T09:16:48.0377926Z * [new tag] ciflow/binaries_wheel/167981 -> ciflow/binaries_wheel/167981 2025-12-04T09:16:48.0379493Z * [new tag] ciflow/dynamo/167695 -> ciflow/dynamo/167695 2025-12-04T09:16:48.0380753Z * [new tag] ciflow/dynamo/168096 -> ciflow/dynamo/168096 2025-12-04T09:16:48.0382167Z * [new tag] ciflow/dynamo/169525 -> ciflow/dynamo/169525 2025-12-04T09:16:48.0383757Z * [new tag] ciflow/h100-cutlass-backend/161938 -> ciflow/h100-cutlass-backend/161938 2025-12-04T09:16:48.0385144Z * [new tag] ciflow/h100-cutlass-backend/161940 -> ciflow/h100-cutlass-backend/161940 2025-12-04T09:16:48.0386763Z * [new tag] ciflow/h100-distributed/168923 -> ciflow/h100-distributed/168923 2025-12-04T09:16:48.0388278Z * [new tag] ciflow/h100-symm-mem/167552 -> ciflow/h100-symm-mem/167552 2025-12-04T09:16:48.0389509Z * [new tag] ciflow/h100-symm-mem/168129 -> ciflow/h100-symm-mem/168129 2025-12-04T09:16:48.0390742Z * [new tag] ciflow/h100-symm-mem/168917 -> ciflow/h100-symm-mem/168917 2025-12-04T09:16:48.0392326Z * [new tag] ciflow/h100-symm-mem/169156 -> ciflow/h100-symm-mem/169156 2025-12-04T09:16:48.0393602Z * [new tag] ciflow/h100-symm-mem/169200 -> ciflow/h100-symm-mem/169200 2025-12-04T09:16:48.0394866Z * [new tag] ciflow/h100-symm-mem/169216 -> ciflow/h100-symm-mem/169216 2025-12-04T09:16:48.0396129Z * [new tag] ciflow/h100-symm-mem/169338 -> ciflow/h100-symm-mem/169338 2025-12-04T09:16:48.0397510Z * [new tag] ciflow/h100-symm-mem/169355 -> ciflow/h100-symm-mem/169355 2025-12-04T09:16:48.0398797Z * [new tag] ciflow/h100-symm-mem/169543 -> ciflow/h100-symm-mem/169543 2025-12-04T09:16:48.0400288Z * [new tag] ciflow/h100/115316 -> ciflow/h100/115316 2025-12-04T09:16:48.0401516Z * [new tag] ciflow/h100/160685 -> ciflow/h100/160685 2025-12-04T09:16:48.0402779Z * [new tag] ciflow/h100/160729 -> ciflow/h100/160729 2025-12-04T09:16:48.0404038Z * [new tag] ciflow/h100/161607 -> ciflow/h100/161607 2025-12-04T09:16:48.0405270Z * [new tag] ciflow/h100/161938 -> ciflow/h100/161938 2025-12-04T09:16:48.0406652Z * [new tag] ciflow/h100/167207 -> ciflow/h100/167207 2025-12-04T09:16:48.0407851Z * [new tag] ciflow/h100/167989 -> ciflow/h100/167989 2025-12-04T09:16:48.0409151Z * [new tag] ciflow/h100/168096 -> ciflow/h100/168096 2025-12-04T09:16:48.0410438Z * [new tag] ciflow/h100/168175 -> ciflow/h100/168175 2025-12-04T09:16:48.0411660Z * [new tag] ciflow/h100/168195 -> ciflow/h100/168195 2025-12-04T09:16:48.0412911Z * [new tag] ciflow/h100/168980 -> ciflow/h100/168980 2025-12-04T09:16:48.0414534Z * [new tag] ciflow/h100/169200 -> ciflow/h100/169200 2025-12-04T09:16:48.0416222Z * [new tag] ciflow/h100/169216 -> ciflow/h100/169216 2025-12-04T09:16:48.0417753Z * [new tag] ciflow/h100/169380 -> ciflow/h100/169380 2025-12-04T09:16:48.0419101Z * [new tag] ciflow/h100/169412 -> ciflow/h100/169412 2025-12-04T09:16:48.0420400Z * [new tag] ciflow/h100/169470 -> ciflow/h100/169470 2025-12-04T09:16:48.0421702Z * [new tag] ciflow/h100/169471 -> ciflow/h100/169471 2025-12-04T09:16:48.0423011Z * [new tag] ciflow/h100/169472 -> ciflow/h100/169472 2025-12-04T09:16:48.0424449Z * [new tag] ciflow/h100/169514 -> ciflow/h100/169514 2025-12-04T09:16:48.0426063Z * [new tag] ciflow/inductor-cu126/168096 -> ciflow/inductor-cu126/168096 2025-12-04T09:16:48.0427972Z * [new tag] ciflow/inductor-micro-benchmark-cpu-x86/168096 -> ciflow/inductor-micro-benchmark-cpu-x86/168096 2025-12-04T09:16:48.0429531Z * [new tag] ciflow/inductor-micro-benchmark/166165 -> ciflow/inductor-micro-benchmark/166165 2025-12-04T09:16:48.0430848Z * [new tag] ciflow/inductor-micro-benchmark/168096 -> ciflow/inductor-micro-benchmark/168096 2025-12-04T09:16:48.0432413Z * [new tag] ciflow/inductor-perf-compare/168096 -> ciflow/inductor-perf-compare/168096 2025-12-04T09:16:48.0434306Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168073 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168073 2025-12-04T09:16:48.0435587Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168096 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168096 2025-12-04T09:16:48.0437358Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi300/169024 2025-12-04T09:16:48.0439043Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi355/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi355/169024 2025-12-04T09:16:48.0440484Z * [new tag] ciflow/inductor-perf-test-nightly/168096 -> ciflow/inductor-perf-test-nightly/168096 2025-12-04T09:16:48.0442010Z * [new tag] ciflow/inductor-periodic/168096 -> ciflow/inductor-periodic/168096 2025-12-04T09:16:48.0443263Z * [new tag] ciflow/inductor-periodic/169024 -> ciflow/inductor-periodic/169024 2025-12-04T09:16:48.0444631Z * [new tag] ciflow/inductor-periodic/169425 -> ciflow/inductor-periodic/169425 2025-12-04T09:16:48.0446280Z * [new tag] ciflow/inductor-rocm-mi200/165545 -> ciflow/inductor-rocm-mi200/165545 2025-12-04T09:16:48.0447716Z * [new tag] ciflow/inductor-rocm-mi200/165997 -> ciflow/inductor-rocm-mi200/165997 2025-12-04T09:16:48.0449091Z * [new tag] ciflow/inductor-rocm-mi200/168096 -> ciflow/inductor-rocm-mi200/168096 2025-12-04T09:16:48.0450480Z * [new tag] ciflow/inductor-rocm-mi200/169063 -> ciflow/inductor-rocm-mi200/169063 2025-12-04T09:16:48.0451717Z * [new tag] ciflow/inductor-rocm-mi200/169425 -> ciflow/inductor-rocm-mi200/169425 2025-12-04T09:16:48.0453396Z * [new tag] ciflow/inductor-rocm-mi300/165545 -> ciflow/inductor-rocm-mi300/165545 2025-12-04T09:16:48.0454311Z * [new tag] ciflow/inductor-rocm-mi300/168096 -> ciflow/inductor-rocm-mi300/168096 2025-12-04T09:16:48.0455685Z * [new tag] ciflow/inductor-rocm-mi300/169063 -> ciflow/inductor-rocm-mi300/169063 2025-12-04T09:16:48.0457045Z * [new tag] ciflow/inductor-rocm-mi300/169425 -> ciflow/inductor-rocm-mi300/169425 2025-12-04T09:16:48.0458638Z * [new tag] ciflow/inductor-rocm/162052 -> ciflow/inductor-rocm/162052 2025-12-04T09:16:48.0460374Z * [new tag] ciflow/inductor-rocm/168971 -> ciflow/inductor-rocm/168971 2025-12-04T09:16:48.0461961Z * [new tag] ciflow/inductor-windows/168096 -> ciflow/inductor-windows/168096 2025-12-04T09:16:48.0463395Z * [new tag] ciflow/inductor/144542 -> ciflow/inductor/144542 2025-12-04T09:16:48.0464643Z * [new tag] ciflow/inductor/146506 -> ciflow/inductor/146506 2025-12-04T09:16:48.0465892Z * [new tag] ciflow/inductor/147990 -> ciflow/inductor/147990 2025-12-04T09:16:48.0467235Z * [new tag] ciflow/inductor/148294 -> ciflow/inductor/148294 2025-12-04T09:16:48.0468484Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-12-04T09:16:48.0469753Z * [new tag] ciflow/inductor/157149 -> ciflow/inductor/157149 2025-12-04T09:16:48.0470988Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-12-04T09:16:48.0472193Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-12-04T09:16:48.0473446Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-12-04T09:16:48.0474655Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-12-04T09:16:48.0475926Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-12-04T09:16:48.0477552Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-12-04T09:16:48.0479240Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-12-04T09:16:48.0480764Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-12-04T09:16:48.0482143Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-12-04T09:16:48.0483403Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-12-04T09:16:48.0484675Z * [new tag] ciflow/inductor/162052 -> ciflow/inductor/162052 2025-12-04T09:16:48.0485986Z * [new tag] ciflow/inductor/162275 -> ciflow/inductor/162275 2025-12-04T09:16:48.0487373Z * [new tag] ciflow/inductor/162795 -> ciflow/inductor/162795 2025-12-04T09:16:48.0489010Z * [new tag] ciflow/inductor/163245 -> ciflow/inductor/163245 2025-12-04T09:16:48.0490321Z * [new tag] ciflow/inductor/163335 -> ciflow/inductor/163335 2025-12-04T09:16:48.0491621Z * [new tag] ciflow/inductor/163503 -> ciflow/inductor/163503 2025-12-04T09:16:48.0492960Z * [new tag] ciflow/inductor/163942 -> ciflow/inductor/163942 2025-12-04T09:16:48.0494338Z * [new tag] ciflow/inductor/165270 -> ciflow/inductor/165270 2025-12-04T09:16:48.0495647Z * [new tag] ciflow/inductor/165274 -> ciflow/inductor/165274 2025-12-04T09:16:48.0497080Z * [new tag] ciflow/inductor/165322 -> ciflow/inductor/165322 2025-12-04T09:16:48.0498967Z * [new tag] ciflow/inductor/165597 -> ciflow/inductor/165597 2025-12-04T09:16:48.0500285Z * [new tag] ciflow/inductor/166063 -> ciflow/inductor/166063 2025-12-04T09:16:48.0501583Z * [new tag] ciflow/inductor/166075 -> ciflow/inductor/166075 2025-12-04T09:16:48.0502987Z * [new tag] ciflow/inductor/166165 -> ciflow/inductor/166165 2025-12-04T09:16:48.0504461Z * [new tag] ciflow/inductor/166254 -> ciflow/inductor/166254 2025-12-04T09:16:48.0505733Z * [new tag] ciflow/inductor/166483 -> ciflow/inductor/166483 2025-12-04T09:16:48.0507015Z * [new tag] ciflow/inductor/166494 -> ciflow/inductor/166494 2025-12-04T09:16:48.0508355Z * [new tag] ciflow/inductor/166545 -> ciflow/inductor/166545 2025-12-04T09:16:48.0509698Z * [new tag] ciflow/inductor/166788 -> ciflow/inductor/166788 2025-12-04T09:16:48.0511191Z * [new tag] ciflow/inductor/166846 -> ciflow/inductor/166846 2025-12-04T09:16:48.0512515Z * [new tag] ciflow/inductor/167300 -> ciflow/inductor/167300 2025-12-04T09:16:48.0513864Z * [new tag] ciflow/inductor/167407 -> ciflow/inductor/167407 2025-12-04T09:16:48.0515327Z * [new tag] ciflow/inductor/167536 -> ciflow/inductor/167536 2025-12-04T09:16:48.0516642Z * [new tag] ciflow/inductor/167552 -> ciflow/inductor/167552 2025-12-04T09:16:48.0517987Z * [new tag] ciflow/inductor/167555 -> ciflow/inductor/167555 2025-12-04T09:16:48.0519479Z * [new tag] ciflow/inductor/167583 -> ciflow/inductor/167583 2025-12-04T09:16:48.0520771Z * [new tag] ciflow/inductor/167599 -> ciflow/inductor/167599 2025-12-04T09:16:48.0522087Z * [new tag] ciflow/inductor/167647 -> ciflow/inductor/167647 2025-12-04T09:16:48.0523390Z * [new tag] ciflow/inductor/167677 -> ciflow/inductor/167677 2025-12-04T09:16:48.0524771Z * [new tag] ciflow/inductor/167680 -> ciflow/inductor/167680 2025-12-04T09:16:48.0526100Z * [new tag] ciflow/inductor/167695 -> ciflow/inductor/167695 2025-12-04T09:16:48.0527392Z * [new tag] ciflow/inductor/167742 -> ciflow/inductor/167742 2025-12-04T09:16:48.0528845Z * [new tag] ciflow/inductor/167768 -> ciflow/inductor/167768 2025-12-04T09:16:48.0530446Z * [new tag] ciflow/inductor/167773 -> ciflow/inductor/167773 2025-12-04T09:16:48.0531818Z * [new tag] ciflow/inductor/167781 -> ciflow/inductor/167781 2025-12-04T09:16:48.0533074Z * [new tag] ciflow/inductor/167880 -> ciflow/inductor/167880 2025-12-04T09:16:48.0534432Z * [new tag] ciflow/inductor/167887 -> ciflow/inductor/167887 2025-12-04T09:16:48.0536090Z * [new tag] ciflow/inductor/167972 -> ciflow/inductor/167972 2025-12-04T09:16:48.0539782Z * [new tag] ciflow/inductor/167989 -> ciflow/inductor/167989 2025-12-04T09:16:48.0541103Z * [new tag] ciflow/inductor/168002 -> ciflow/inductor/168002 2025-12-04T09:16:48.0542381Z * [new tag] ciflow/inductor/168050 -> ciflow/inductor/168050 2025-12-04T09:16:48.0543777Z * [new tag] ciflow/inductor/168051 -> ciflow/inductor/168051 2025-12-04T09:16:48.0545128Z * [new tag] ciflow/inductor/168052 -> ciflow/inductor/168052 2025-12-04T09:16:48.0546420Z * [new tag] ciflow/inductor/168073 -> ciflow/inductor/168073 2025-12-04T09:16:48.0547736Z * [new tag] ciflow/inductor/168096 -> ciflow/inductor/168096 2025-12-04T09:16:48.0549050Z * [new tag] ciflow/inductor/168114 -> ciflow/inductor/168114 2025-12-04T09:16:48.0550350Z * [new tag] ciflow/inductor/168115 -> ciflow/inductor/168115 2025-12-04T09:16:48.0551666Z * [new tag] ciflow/inductor/168127 -> ciflow/inductor/168127 2025-12-04T09:16:48.0552973Z * [new tag] ciflow/inductor/168129 -> ciflow/inductor/168129 2025-12-04T09:16:48.0554318Z * [new tag] ciflow/inductor/168157 -> ciflow/inductor/168157 2025-12-04T09:16:48.0555765Z * [new tag] ciflow/inductor/168175 -> ciflow/inductor/168175 2025-12-04T09:16:48.0556886Z * [new tag] ciflow/inductor/168185 -> ciflow/inductor/168185 2025-12-04T09:16:48.0558258Z * [new tag] ciflow/inductor/168195 -> ciflow/inductor/168195 2025-12-04T09:16:48.0559497Z * [new tag] ciflow/inductor/168209 -> ciflow/inductor/168209 2025-12-04T09:16:48.0560877Z * [new tag] ciflow/inductor/168266 -> ciflow/inductor/168266 2025-12-04T09:16:48.0562188Z * [new tag] ciflow/inductor/168316 -> ciflow/inductor/168316 2025-12-04T09:16:48.0563678Z * [new tag] ciflow/inductor/168326 -> ciflow/inductor/168326 2025-12-04T09:16:48.0564987Z * [new tag] ciflow/inductor/168368 -> ciflow/inductor/168368 2025-12-04T09:16:48.0566285Z * [new tag] ciflow/inductor/168894 -> ciflow/inductor/168894 2025-12-04T09:16:48.0567679Z * [new tag] ciflow/inductor/168934 -> ciflow/inductor/168934 2025-12-04T09:16:48.0569053Z * [new tag] ciflow/inductor/168939 -> ciflow/inductor/168939 2025-12-04T09:16:48.0570373Z * [new tag] ciflow/inductor/168946 -> ciflow/inductor/168946 2025-12-04T09:16:48.0571659Z * [new tag] ciflow/inductor/168950 -> ciflow/inductor/168950 2025-12-04T09:16:48.0573000Z * [new tag] ciflow/inductor/168951 -> ciflow/inductor/168951 2025-12-04T09:16:48.0574317Z * [new tag] ciflow/inductor/168952 -> ciflow/inductor/168952 2025-12-04T09:16:48.0575677Z * [new tag] ciflow/inductor/168955 -> ciflow/inductor/168955 2025-12-04T09:16:48.0576980Z * [new tag] ciflow/inductor/168971 -> ciflow/inductor/168971 2025-12-04T09:16:48.0578286Z * [new tag] ciflow/inductor/168979 -> ciflow/inductor/168979 2025-12-04T09:16:48.0579609Z * [new tag] ciflow/inductor/168980 -> ciflow/inductor/168980 2025-12-04T09:16:48.0581091Z * [new tag] ciflow/inductor/168983 -> ciflow/inductor/168983 2025-12-04T09:16:48.0582389Z * [new tag] ciflow/inductor/169006 -> ciflow/inductor/169006 2025-12-04T09:16:48.0583769Z * [new tag] ciflow/inductor/169023 -> ciflow/inductor/169023 2025-12-04T09:16:48.0585076Z * [new tag] ciflow/inductor/169024 -> ciflow/inductor/169024 2025-12-04T09:16:48.0586532Z * [new tag] ciflow/inductor/169025 -> ciflow/inductor/169025 2025-12-04T09:16:48.0588234Z * [new tag] ciflow/inductor/169066 -> ciflow/inductor/169066 2025-12-04T09:16:48.0589567Z * [new tag] ciflow/inductor/169091 -> ciflow/inductor/169091 2025-12-04T09:16:48.0590909Z * [new tag] ciflow/inductor/169102 -> ciflow/inductor/169102 2025-12-04T09:16:48.0592274Z * [new tag] ciflow/inductor/169103 -> ciflow/inductor/169103 2025-12-04T09:16:48.0593558Z * [new tag] ciflow/inductor/169121 -> ciflow/inductor/169121 2025-12-04T09:16:48.0594862Z * [new tag] ciflow/inductor/169134 -> ciflow/inductor/169134 2025-12-04T09:16:48.0596210Z * [new tag] ciflow/inductor/169135 -> ciflow/inductor/169135 2025-12-04T09:16:48.0597541Z * [new tag] ciflow/inductor/169141 -> ciflow/inductor/169141 2025-12-04T09:16:48.0598894Z * [new tag] ciflow/inductor/169151 -> ciflow/inductor/169151 2025-12-04T09:16:48.0600347Z * [new tag] ciflow/inductor/169161 -> ciflow/inductor/169161 2025-12-04T09:16:48.0601695Z * [new tag] ciflow/inductor/169167 -> ciflow/inductor/169167 2025-12-04T09:16:48.0603192Z * [new tag] ciflow/inductor/169177 -> ciflow/inductor/169177 2025-12-04T09:16:48.0604743Z * [new tag] ciflow/inductor/169185 -> ciflow/inductor/169185 2025-12-04T09:16:48.0605983Z * [new tag] ciflow/inductor/169196 -> ciflow/inductor/169196 2025-12-04T09:16:48.0607290Z * [new tag] ciflow/inductor/169200 -> ciflow/inductor/169200 2025-12-04T09:16:48.0608839Z * [new tag] ciflow/inductor/169204 -> ciflow/inductor/169204 2025-12-04T09:16:48.0609888Z * [new tag] ciflow/inductor/169216 -> ciflow/inductor/169216 2025-12-04T09:16:48.0611418Z * [new tag] ciflow/inductor/169219 -> ciflow/inductor/169219 2025-12-04T09:16:48.0612736Z * [new tag] ciflow/inductor/169220 -> ciflow/inductor/169220 2025-12-04T09:16:48.0614267Z * [new tag] ciflow/inductor/169230 -> ciflow/inductor/169230 2025-12-04T09:16:48.0615663Z * [new tag] ciflow/inductor/169242 -> ciflow/inductor/169242 2025-12-04T09:16:48.0616986Z * [new tag] ciflow/inductor/169245 -> ciflow/inductor/169245 2025-12-04T09:16:48.0618426Z * [new tag] ciflow/inductor/169260 -> ciflow/inductor/169260 2025-12-04T09:16:48.0619773Z * [new tag] ciflow/inductor/169282 -> ciflow/inductor/169282 2025-12-04T09:16:48.0621071Z * [new tag] ciflow/inductor/169286 -> ciflow/inductor/169286 2025-12-04T09:16:48.0622384Z * [new tag] ciflow/inductor/169299 -> ciflow/inductor/169299 2025-12-04T09:16:48.0623857Z * [new tag] ciflow/inductor/169304 -> ciflow/inductor/169304 2025-12-04T09:16:48.0625582Z * [new tag] ciflow/inductor/169305 -> ciflow/inductor/169305 2025-12-04T09:16:48.0626901Z * [new tag] ciflow/inductor/169308 -> ciflow/inductor/169308 2025-12-04T09:16:48.0628251Z * [new tag] ciflow/inductor/169319 -> ciflow/inductor/169319 2025-12-04T09:16:48.0629575Z * [new tag] ciflow/inductor/169326 -> ciflow/inductor/169326 2025-12-04T09:16:48.0631016Z * [new tag] ciflow/inductor/169332 -> ciflow/inductor/169332 2025-12-04T09:16:48.0632328Z * [new tag] ciflow/inductor/169333 -> ciflow/inductor/169333 2025-12-04T09:16:48.0633863Z * [new tag] ciflow/inductor/169336 -> ciflow/inductor/169336 2025-12-04T09:16:48.0635387Z * [new tag] ciflow/inductor/169340 -> ciflow/inductor/169340 2025-12-04T09:16:48.0637016Z * [new tag] ciflow/inductor/169341 -> ciflow/inductor/169341 2025-12-04T09:16:48.0638400Z * [new tag] ciflow/inductor/169343 -> ciflow/inductor/169343 2025-12-04T09:16:48.0639724Z * [new tag] ciflow/inductor/169346 -> ciflow/inductor/169346 2025-12-04T09:16:48.0641224Z * [new tag] ciflow/inductor/169348 -> ciflow/inductor/169348 2025-12-04T09:16:48.0642668Z * [new tag] ciflow/inductor/169350 -> ciflow/inductor/169350 2025-12-04T09:16:48.0644008Z * [new tag] ciflow/inductor/169355 -> ciflow/inductor/169355 2025-12-04T09:16:48.0645354Z * [new tag] ciflow/inductor/169370 -> ciflow/inductor/169370 2025-12-04T09:16:48.0647002Z * [new tag] ciflow/inductor/169375 -> ciflow/inductor/169375 2025-12-04T09:16:48.0648401Z * [new tag] ciflow/inductor/169389 -> ciflow/inductor/169389 2025-12-04T09:16:48.0649748Z * [new tag] ciflow/inductor/169391 -> ciflow/inductor/169391 2025-12-04T09:16:48.0651105Z * [new tag] ciflow/inductor/169393 -> ciflow/inductor/169393 2025-12-04T09:16:48.0652446Z * [new tag] ciflow/inductor/169399 -> ciflow/inductor/169399 2025-12-04T09:16:48.0653926Z * [new tag] ciflow/inductor/169400 -> ciflow/inductor/169400 2025-12-04T09:16:48.0655249Z * [new tag] ciflow/inductor/169415 -> ciflow/inductor/169415 2025-12-04T09:16:48.0656758Z * [new tag] ciflow/inductor/169417 -> ciflow/inductor/169417 2025-12-04T09:16:48.0658000Z * [new tag] ciflow/inductor/169418 -> ciflow/inductor/169418 2025-12-04T09:16:48.0659495Z * [new tag] ciflow/inductor/169430 -> ciflow/inductor/169430 2025-12-04T09:16:48.0660732Z * [new tag] ciflow/inductor/169432 -> ciflow/inductor/169432 2025-12-04T09:16:48.0662154Z * [new tag] ciflow/inductor/169436 -> ciflow/inductor/169436 2025-12-04T09:16:48.0663685Z * [new tag] ciflow/inductor/169437 -> ciflow/inductor/169437 2025-12-04T09:16:48.0664995Z * [new tag] ciflow/inductor/169438 -> ciflow/inductor/169438 2025-12-04T09:16:48.0666286Z * [new tag] ciflow/inductor/169441 -> ciflow/inductor/169441 2025-12-04T09:16:48.0667610Z * [new tag] ciflow/inductor/169446 -> ciflow/inductor/169446 2025-12-04T09:16:48.0669162Z * [new tag] ciflow/inductor/169447 -> ciflow/inductor/169447 2025-12-04T09:16:48.0670468Z * [new tag] ciflow/inductor/169452 -> ciflow/inductor/169452 2025-12-04T09:16:48.0671972Z * [new tag] ciflow/inductor/169455 -> ciflow/inductor/169455 2025-12-04T09:16:48.0673317Z * [new tag] ciflow/inductor/169459 -> ciflow/inductor/169459 2025-12-04T09:16:48.0674818Z * [new tag] ciflow/inductor/169463 -> ciflow/inductor/169463 2025-12-04T09:16:48.0676343Z * [new tag] ciflow/inductor/169476 -> ciflow/inductor/169476 2025-12-04T09:16:48.0678147Z * [new tag] ciflow/inductor/169485 -> ciflow/inductor/169485 2025-12-04T09:16:48.0679467Z * [new tag] ciflow/inductor/169493 -> ciflow/inductor/169493 2025-12-04T09:16:48.0680861Z * [new tag] ciflow/inductor/169496 -> ciflow/inductor/169496 2025-12-04T09:16:48.0682184Z * [new tag] ciflow/inductor/169497 -> ciflow/inductor/169497 2025-12-04T09:16:48.0683505Z * [new tag] ciflow/inductor/169503 -> ciflow/inductor/169503 2025-12-04T09:16:48.0684828Z * [new tag] ciflow/inductor/169504 -> ciflow/inductor/169504 2025-12-04T09:16:48.0686590Z * [new tag] ciflow/inductor/169505 -> ciflow/inductor/169505 2025-12-04T09:16:48.0688610Z * [new tag] ciflow/inductor/169508 -> ciflow/inductor/169508 2025-12-04T09:16:48.0689953Z * [new tag] ciflow/inductor/169509 -> ciflow/inductor/169509 2025-12-04T09:16:48.0691380Z * [new tag] ciflow/inductor/169513 -> ciflow/inductor/169513 2025-12-04T09:16:48.0692669Z * [new tag] ciflow/inductor/169514 -> ciflow/inductor/169514 2025-12-04T09:16:48.0693941Z * [new tag] ciflow/inductor/169515 -> ciflow/inductor/169515 2025-12-04T09:16:48.0695552Z * [new tag] ciflow/inductor/169517 -> ciflow/inductor/169517 2025-12-04T09:16:48.0696621Z * [new tag] ciflow/inductor/169519 -> ciflow/inductor/169519 2025-12-04T09:16:48.0698285Z * [new tag] ciflow/inductor/169520 -> ciflow/inductor/169520 2025-12-04T09:16:48.0699397Z * [new tag] ciflow/inductor/169521 -> ciflow/inductor/169521 2025-12-04T09:16:48.0700734Z * [new tag] ciflow/inductor/169524 -> ciflow/inductor/169524 2025-12-04T09:16:48.0702231Z * [new tag] ciflow/inductor/169527 -> ciflow/inductor/169527 2025-12-04T09:16:48.0703324Z * [new tag] ciflow/inductor/169528 -> ciflow/inductor/169528 2025-12-04T09:16:48.0705494Z * [new tag] ciflow/inductor/169532 -> ciflow/inductor/169532 2025-12-04T09:16:48.0706160Z * [new tag] ciflow/inductor/169535 -> ciflow/inductor/169535 2025-12-04T09:16:48.0707828Z * [new tag] ciflow/inductor/169536 -> ciflow/inductor/169536 2025-12-04T09:16:48.0708974Z * [new tag] ciflow/inductor/169547 -> ciflow/inductor/169547 2025-12-04T09:16:48.0710173Z * [new tag] ciflow/inductor/169548 -> ciflow/inductor/169548 2025-12-04T09:16:48.0712071Z * [new tag] ciflow/inductor/169549 -> ciflow/inductor/169549 2025-12-04T09:16:48.0713025Z * [new tag] ciflow/inductor/169551 -> ciflow/inductor/169551 2025-12-04T09:16:48.0714306Z * [new tag] ciflow/inductor/169552 -> ciflow/inductor/169552 2025-12-04T09:16:48.0715843Z * [new tag] ciflow/inductor/169553 -> ciflow/inductor/169553 2025-12-04T09:16:48.0716990Z * [new tag] ciflow/inductor/169557 -> ciflow/inductor/169557 2025-12-04T09:16:48.0719183Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-12-04T09:16:48.0720141Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-12-04T09:16:48.0722141Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-12-04T09:16:48.0723306Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-12-04T09:16:48.0724586Z * [new tag] ciflow/linux-aarch64/166075 -> ciflow/linux-aarch64/166075 2025-12-04T09:16:48.0725796Z * [new tag] ciflow/linux-aarch64/166876 -> ciflow/linux-aarch64/166876 2025-12-04T09:16:48.0727008Z * [new tag] ciflow/linux-aarch64/167981 -> ciflow/linux-aarch64/167981 2025-12-04T09:16:48.0729072Z * [new tag] ciflow/mps/166254 -> ciflow/mps/166254 2025-12-04T09:16:48.0729946Z * [new tag] ciflow/mps/169017 -> ciflow/mps/169017 2025-12-04T09:16:48.0731222Z * [new tag] ciflow/mps/169372 -> ciflow/mps/169372 2025-12-04T09:16:48.0732755Z * [new tag] ciflow/mps/169478 -> ciflow/mps/169478 2025-12-04T09:16:48.0734528Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-12-04T09:16:48.0735499Z * [new tag] ciflow/op-benchmark/166075 -> ciflow/op-benchmark/166075 2025-12-04T09:16:48.0736522Z * [new tag] ciflow/op-benchmark/169544 -> ciflow/op-benchmark/169544 2025-12-04T09:16:48.0738359Z * [new tag] ciflow/periodic-rocm-mi200/165997 -> ciflow/periodic-rocm-mi200/165997 2025-12-04T09:16:48.0739733Z * [new tag] ciflow/periodic-rocm-mi200/166517 -> ciflow/periodic-rocm-mi200/166517 2025-12-04T09:16:48.0740739Z * [new tag] ciflow/periodic-rocm-mi200/169063 -> ciflow/periodic-rocm-mi200/169063 2025-12-04T09:16:48.0742200Z * [new tag] ciflow/periodic-rocm-mi200/169425 -> ciflow/periodic-rocm-mi200/169425 2025-12-04T09:16:48.0743793Z * [new tag] ciflow/periodic-rocm-mi300/166517 -> ciflow/periodic-rocm-mi300/166517 2025-12-04T09:16:48.0744807Z * [new tag] ciflow/periodic-rocm-mi300/169063 -> ciflow/periodic-rocm-mi300/169063 2025-12-04T09:16:48.0746243Z * [new tag] ciflow/periodic-rocm-mi300/169425 -> ciflow/periodic-rocm-mi300/169425 2025-12-04T09:16:48.0748005Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-12-04T09:16:48.0749246Z * [new tag] ciflow/periodic/167207 -> ciflow/periodic/167207 2025-12-04T09:16:48.0750599Z * [new tag] ciflow/periodic/167978 -> ciflow/periodic/167978 2025-12-04T09:16:48.0751857Z * [new tag] ciflow/periodic/168096 -> ciflow/periodic/168096 2025-12-04T09:16:48.0753086Z * [new tag] ciflow/periodic/169286 -> ciflow/periodic/169286 2025-12-04T09:16:48.0754643Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-12-04T09:16:48.0755990Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-12-04T09:16:48.0757663Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-12-04T09:16:48.0759040Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-12-04T09:16:48.0761101Z * [new tag] ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T09:16:48.0762654Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-12-04T09:16:48.0764527Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-12-04T09:16:48.0766045Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-12-04T09:16:48.0767556Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-12-04T09:16:48.0769175Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-12-04T09:16:48.0770932Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-12-04T09:16:48.0772587Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-12-04T09:16:48.0774043Z * [new tag] ciflow/pull/167207 -> ciflow/pull/167207 2025-12-04T09:16:48.0775876Z * [new tag] ciflow/quantization-periodic/169207 -> ciflow/quantization-periodic/169207 2025-12-04T09:16:48.0777342Z * [new tag] ciflow/rocm-mi200/165545 -> ciflow/rocm-mi200/165545 2025-12-04T09:16:48.0778560Z * [new tag] ciflow/rocm-mi200/165997 -> ciflow/rocm-mi200/165997 2025-12-04T09:16:48.0779796Z * [new tag] ciflow/rocm-mi200/168096 -> ciflow/rocm-mi200/168096 2025-12-04T09:16:48.0781208Z * [new tag] ciflow/rocm-mi200/168275 -> ciflow/rocm-mi200/168275 2025-12-04T09:16:48.0782448Z * [new tag] ciflow/rocm-mi200/169063 -> ciflow/rocm-mi200/169063 2025-12-04T09:16:48.0783880Z * [new tag] ciflow/rocm-mi200/169356 -> ciflow/rocm-mi200/169356 2025-12-04T09:16:48.0784956Z * [new tag] ciflow/rocm-mi200/169425 -> ciflow/rocm-mi200/169425 2025-12-04T09:16:48.0786666Z * [new tag] ciflow/rocm-mi300/165545 -> ciflow/rocm-mi300/165545 2025-12-04T09:16:48.0788136Z * [new tag] ciflow/rocm-mi300/167157 -> ciflow/rocm-mi300/167157 2025-12-04T09:16:48.0789357Z * [new tag] ciflow/rocm-mi300/168096 -> ciflow/rocm-mi300/168096 2025-12-04T09:16:48.0790649Z * [new tag] ciflow/rocm-mi300/169063 -> ciflow/rocm-mi300/169063 2025-12-04T09:16:48.0791953Z * [new tag] ciflow/rocm-mi300/169425 -> ciflow/rocm-mi300/169425 2025-12-04T09:16:48.0793758Z * [new tag] ciflow/rocm-mi355/167157 -> ciflow/rocm-mi355/167157 2025-12-04T09:16:48.0795033Z * [new tag] ciflow/rocm-mi355/168275 -> ciflow/rocm-mi355/168275 2025-12-04T09:16:48.0796379Z * [new tag] ciflow/rocm-mi355/169425 -> ciflow/rocm-mi355/169425 2025-12-04T09:16:48.0797989Z * [new tag] ciflow/rocm-navi31/168275 -> ciflow/rocm-navi31/168275 2025-12-04T09:16:48.0799079Z * [new tag] ciflow/rocm-navi31/169425 -> ciflow/rocm-navi31/169425 2025-12-04T09:16:48.0800677Z * [new tag] ciflow/rocm/115316 -> ciflow/rocm/115316 2025-12-04T09:16:48.0801917Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-12-04T09:16:48.0803212Z * [new tag] ciflow/rocm/160685 -> ciflow/rocm/160685 2025-12-04T09:16:48.0804271Z * [new tag] ciflow/rocm/161607 -> ciflow/rocm/161607 2025-12-04T09:16:48.0805640Z * [new tag] ciflow/rocm/162052 -> ciflow/rocm/162052 2025-12-04T09:16:48.0806913Z * [new tag] ciflow/rocm/165997 -> ciflow/rocm/165997 2025-12-04T09:16:48.0808402Z * [new tag] ciflow/rocm/166165 -> ciflow/rocm/166165 2025-12-04T09:16:48.0809332Z * [new tag] ciflow/rocm/166517 -> ciflow/rocm/166517 2025-12-04T09:16:48.0810683Z * [new tag] ciflow/rocm/167207 -> ciflow/rocm/167207 2025-12-04T09:16:48.0811920Z * [new tag] ciflow/rocm/167536 -> ciflow/rocm/167536 2025-12-04T09:16:48.0813138Z * [new tag] ciflow/rocm/167781 -> ciflow/rocm/167781 2025-12-04T09:16:48.0814757Z * [new tag] ciflow/rocm/167989 -> ciflow/rocm/167989 2025-12-04T09:16:48.0816454Z * [new tag] ciflow/rocm/168073 -> ciflow/rocm/168073 2025-12-04T09:16:48.0817993Z * [new tag] ciflow/rocm/168195 -> ciflow/rocm/168195 2025-12-04T09:16:48.0819420Z * [new tag] ciflow/rocm/168939 -> ciflow/rocm/168939 2025-12-04T09:16:48.0820700Z * [new tag] ciflow/rocm/168971 -> ciflow/rocm/168971 2025-12-04T09:16:48.0821985Z * [new tag] ciflow/rocm/169024 -> ciflow/rocm/169024 2025-12-04T09:16:48.0823259Z * [new tag] ciflow/rocm/169200 -> ciflow/rocm/169200 2025-12-04T09:16:48.0824614Z * [new tag] ciflow/rocm/169216 -> ciflow/rocm/169216 2025-12-04T09:16:48.0825883Z * [new tag] ciflow/rocm/169312 -> ciflow/rocm/169312 2025-12-04T09:16:48.0827219Z * [new tag] ciflow/rocm/169380 -> ciflow/rocm/169380 2025-12-04T09:16:48.0828541Z * [new tag] ciflow/rocm/169427 -> ciflow/rocm/169427 2025-12-04T09:16:48.0829921Z * [new tag] ciflow/rocm/169455 -> ciflow/rocm/169455 2025-12-04T09:16:48.0831209Z * [new tag] ciflow/rocm/169470 -> ciflow/rocm/169470 2025-12-04T09:16:48.0832549Z * [new tag] ciflow/rocm/169471 -> ciflow/rocm/169471 2025-12-04T09:16:48.0833876Z * [new tag] ciflow/rocm/169472 -> ciflow/rocm/169472 2025-12-04T09:16:48.0835170Z * [new tag] ciflow/rocm/169514 -> ciflow/rocm/169514 2025-12-04T09:16:48.0837472Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-12-04T09:16:48.0838882Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-12-04T09:16:48.0840719Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-12-04T09:16:48.0841632Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-12-04T09:16:48.0843043Z * [new tag] ciflow/slow/167207 -> ciflow/slow/167207 2025-12-04T09:16:48.0844269Z * [new tag] ciflow/slow/168050 -> ciflow/slow/168050 2025-12-04T09:16:48.0845722Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-12-04T09:16:48.0847238Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-12-04T09:16:48.0849232Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-12-04T09:16:48.0851009Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-12-04T09:16:48.0852765Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-12-04T09:16:48.0854268Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-12-04T09:16:48.0855711Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-12-04T09:16:48.0857163Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-12-04T09:16:48.0859044Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-12-04T09:16:48.0860106Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-12-04T09:16:48.0861823Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-12-04T09:16:48.0863242Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-12-04T09:16:48.0864734Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-12-04T09:16:48.0866281Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-12-04T09:16:48.0868212Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-12-04T09:16:48.0869563Z * [new tag] ciflow/torchbench/168175 -> ciflow/torchbench/168175 2025-12-04T09:16:48.0871053Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-12-04T09:16:48.0872334Z * [new tag] ciflow/trunk/157149 -> ciflow/trunk/157149 2025-12-04T09:16:48.0873576Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-12-04T09:16:48.0874810Z * [new tag] ciflow/trunk/159718 -> ciflow/trunk/159718 2025-12-04T09:16:48.0876038Z * [new tag] ciflow/trunk/160685 -> ciflow/trunk/160685 2025-12-04T09:16:48.0877303Z * [new tag] ciflow/trunk/160729 -> ciflow/trunk/160729 2025-12-04T09:16:48.0878536Z * [new tag] ciflow/trunk/162275 -> ciflow/trunk/162275 2025-12-04T09:16:48.0879867Z * [new tag] ciflow/trunk/162795 -> ciflow/trunk/162795 2025-12-04T09:16:48.0881628Z * [new tag] ciflow/trunk/163245 -> ciflow/trunk/163245 2025-12-04T09:16:48.0882870Z * [new tag] ciflow/trunk/163942 -> ciflow/trunk/163942 2025-12-04T09:16:48.0884127Z * [new tag] ciflow/trunk/165274 -> ciflow/trunk/165274 2025-12-04T09:16:48.0885877Z * [new tag] ciflow/trunk/165483 -> ciflow/trunk/165483 2025-12-04T09:16:48.0887605Z * [new tag] ciflow/trunk/165728 -> ciflow/trunk/165728 2025-12-04T09:16:48.0889220Z * [new tag] ciflow/trunk/165922 -> ciflow/trunk/165922 2025-12-04T09:16:48.0890596Z * [new tag] ciflow/trunk/166075 -> ciflow/trunk/166075 2025-12-04T09:16:48.0891904Z * [new tag] ciflow/trunk/166165 -> ciflow/trunk/166165 2025-12-04T09:16:48.0893209Z * [new tag] ciflow/trunk/166829 -> ciflow/trunk/166829 2025-12-04T09:16:48.0894699Z * [new tag] ciflow/trunk/166843 -> ciflow/trunk/166843 2025-12-04T09:16:48.0895995Z * [new tag] ciflow/trunk/166876 -> ciflow/trunk/166876 2025-12-04T09:16:48.0897310Z * [new tag] ciflow/trunk/167207 -> ciflow/trunk/167207 2025-12-04T09:16:48.0898627Z * [new tag] ciflow/trunk/167536 -> ciflow/trunk/167536 2025-12-04T09:16:48.0899956Z * [new tag] ciflow/trunk/167552 -> ciflow/trunk/167552 2025-12-04T09:16:48.0901271Z * [new tag] ciflow/trunk/167555 -> ciflow/trunk/167555 2025-12-04T09:16:48.0902632Z * [new tag] ciflow/trunk/167599 -> ciflow/trunk/167599 2025-12-04T09:16:48.0903996Z * [new tag] ciflow/trunk/167659 -> ciflow/trunk/167659 2025-12-04T09:16:48.0905425Z * [new tag] ciflow/trunk/167672 -> ciflow/trunk/167672 2025-12-04T09:16:48.0906776Z * [new tag] ciflow/trunk/167742 -> ciflow/trunk/167742 2025-12-04T09:16:48.0908136Z * [new tag] ciflow/trunk/167781 -> ciflow/trunk/167781 2025-12-04T09:16:48.0909679Z * [new tag] ciflow/trunk/167837 -> ciflow/trunk/167837 2025-12-04T09:16:48.0910964Z * [new tag] ciflow/trunk/167887 -> ciflow/trunk/167887 2025-12-04T09:16:48.0912272Z * [new tag] ciflow/trunk/167978 -> ciflow/trunk/167978 2025-12-04T09:16:48.0913699Z * [new tag] ciflow/trunk/168050 -> ciflow/trunk/168050 2025-12-04T09:16:48.0914981Z * [new tag] ciflow/trunk/168051 -> ciflow/trunk/168051 2025-12-04T09:16:48.0916291Z * [new tag] ciflow/trunk/168096 -> ciflow/trunk/168096 2025-12-04T09:16:48.0917608Z * [new tag] ciflow/trunk/168127 -> ciflow/trunk/168127 2025-12-04T09:16:48.0918959Z * [new tag] ciflow/trunk/168157 -> ciflow/trunk/168157 2025-12-04T09:16:48.0920338Z * [new tag] ciflow/trunk/168175 -> ciflow/trunk/168175 2025-12-04T09:16:48.0921616Z * [new tag] ciflow/trunk/168209 -> ciflow/trunk/168209 2025-12-04T09:16:48.0923085Z * [new tag] ciflow/trunk/168213 -> ciflow/trunk/168213 2025-12-04T09:16:48.0924619Z * [new tag] ciflow/trunk/168226 -> ciflow/trunk/168226 2025-12-04T09:16:48.0925996Z * [new tag] ciflow/trunk/168262 -> ciflow/trunk/168262 2025-12-04T09:16:48.0927298Z * [new tag] ciflow/trunk/168275 -> ciflow/trunk/168275 2025-12-04T09:16:48.0928927Z * [new tag] ciflow/trunk/168328 -> ciflow/trunk/168328 2025-12-04T09:16:48.0930222Z * [new tag] ciflow/trunk/168368 -> ciflow/trunk/168368 2025-12-04T09:16:48.0931546Z * [new tag] ciflow/trunk/168917 -> ciflow/trunk/168917 2025-12-04T09:16:48.0932892Z * [new tag] ciflow/trunk/168933 -> ciflow/trunk/168933 2025-12-04T09:16:48.0934339Z * [new tag] ciflow/trunk/168941 -> ciflow/trunk/168941 2025-12-04T09:16:48.0935973Z * [new tag] ciflow/trunk/168955 -> ciflow/trunk/168955 2025-12-04T09:16:48.0939658Z * [new tag] ciflow/trunk/168980 -> ciflow/trunk/168980 2025-12-04T09:16:48.0941166Z * [new tag] ciflow/trunk/169004 -> ciflow/trunk/169004 2025-12-04T09:16:48.0942511Z * [new tag] ciflow/trunk/169006 -> ciflow/trunk/169006 2025-12-04T09:16:48.0943796Z * [new tag] ciflow/trunk/169023 -> ciflow/trunk/169023 2025-12-04T09:16:48.0952791Z * [new tag] ciflow/trunk/169025 -> ciflow/trunk/169025 2025-12-04T09:16:48.0952993Z * [new tag] ciflow/trunk/169048 -> ciflow/trunk/169048 2025-12-04T09:16:48.0953166Z * [new tag] ciflow/trunk/169066 -> ciflow/trunk/169066 2025-12-04T09:16:48.0953332Z * [new tag] ciflow/trunk/169091 -> ciflow/trunk/169091 2025-12-04T09:16:48.0953499Z * [new tag] ciflow/trunk/169102 -> ciflow/trunk/169102 2025-12-04T09:16:48.0953670Z * [new tag] ciflow/trunk/169103 -> ciflow/trunk/169103 2025-12-04T09:16:48.0953835Z * [new tag] ciflow/trunk/169125 -> ciflow/trunk/169125 2025-12-04T09:16:48.0954677Z * [new tag] ciflow/trunk/169139 -> ciflow/trunk/169139 2025-12-04T09:16:48.0956269Z * [new tag] ciflow/trunk/169148 -> ciflow/trunk/169148 2025-12-04T09:16:48.0957535Z * [new tag] ciflow/trunk/169151 -> ciflow/trunk/169151 2025-12-04T09:16:48.0958887Z * [new tag] ciflow/trunk/169156 -> ciflow/trunk/169156 2025-12-04T09:16:48.0960355Z * [new tag] ciflow/trunk/169176 -> ciflow/trunk/169176 2025-12-04T09:16:48.0961759Z * [new tag] ciflow/trunk/169204 -> ciflow/trunk/169204 2025-12-04T09:16:48.0963000Z * [new tag] ciflow/trunk/169207 -> ciflow/trunk/169207 2025-12-04T09:16:48.0964338Z * [new tag] ciflow/trunk/169211 -> ciflow/trunk/169211 2025-12-04T09:16:48.0965869Z * [new tag] ciflow/trunk/169231 -> ciflow/trunk/169231 2025-12-04T09:16:48.0967518Z * [new tag] ciflow/trunk/169260 -> ciflow/trunk/169260 2025-12-04T09:16:48.0968995Z * [new tag] ciflow/trunk/169271 -> ciflow/trunk/169271 2025-12-04T09:16:48.0970329Z * [new tag] ciflow/trunk/169280 -> ciflow/trunk/169280 2025-12-04T09:16:48.0971708Z * [new tag] ciflow/trunk/169281 -> ciflow/trunk/169281 2025-12-04T09:16:48.0973577Z * [new tag] ciflow/trunk/169286 -> ciflow/trunk/169286 2025-12-04T09:16:48.0974995Z * [new tag] ciflow/trunk/169293 -> ciflow/trunk/169293 2025-12-04T09:16:48.0976359Z * [new tag] ciflow/trunk/169296 -> ciflow/trunk/169296 2025-12-04T09:16:48.0977674Z * [new tag] ciflow/trunk/169304 -> ciflow/trunk/169304 2025-12-04T09:16:48.0979001Z * [new tag] ciflow/trunk/169305 -> ciflow/trunk/169305 2025-12-04T09:16:48.0980318Z * [new tag] ciflow/trunk/169312 -> ciflow/trunk/169312 2025-12-04T09:16:48.0981965Z * [new tag] ciflow/trunk/169328 -> ciflow/trunk/169328 2025-12-04T09:16:48.0983324Z * [new tag] ciflow/trunk/169343 -> ciflow/trunk/169343 2025-12-04T09:16:48.0984584Z * [new tag] ciflow/trunk/169355 -> ciflow/trunk/169355 2025-12-04T09:16:48.0985915Z * [new tag] ciflow/trunk/169370 -> ciflow/trunk/169370 2025-12-04T09:16:48.0987402Z * [new tag] ciflow/trunk/169379 -> ciflow/trunk/169379 2025-12-04T09:16:48.0988822Z * [new tag] ciflow/trunk/169380 -> ciflow/trunk/169380 2025-12-04T09:16:48.0990139Z * [new tag] ciflow/trunk/169385 -> ciflow/trunk/169385 2025-12-04T09:16:48.0991453Z * [new tag] ciflow/trunk/169387 -> ciflow/trunk/169387 2025-12-04T09:16:48.0992961Z * [new tag] ciflow/trunk/169410 -> ciflow/trunk/169410 2025-12-04T09:16:48.0994290Z * [new tag] ciflow/trunk/169412 -> ciflow/trunk/169412 2025-12-04T09:16:48.0995639Z * [new tag] ciflow/trunk/169418 -> ciflow/trunk/169418 2025-12-04T09:16:48.0996949Z * [new tag] ciflow/trunk/169423 -> ciflow/trunk/169423 2025-12-04T09:16:48.0998268Z * [new tag] ciflow/trunk/169427 -> ciflow/trunk/169427 2025-12-04T09:16:48.0999608Z * [new tag] ciflow/trunk/169430 -> ciflow/trunk/169430 2025-12-04T09:16:48.1000930Z * [new tag] ciflow/trunk/169437 -> ciflow/trunk/169437 2025-12-04T09:16:48.1002250Z * [new tag] ciflow/trunk/169442 -> ciflow/trunk/169442 2025-12-04T09:16:48.1003580Z * [new tag] ciflow/trunk/169452 -> ciflow/trunk/169452 2025-12-04T09:16:48.1004938Z * [new tag] ciflow/trunk/169454 -> ciflow/trunk/169454 2025-12-04T09:16:48.1006255Z * [new tag] ciflow/trunk/169459 -> ciflow/trunk/169459 2025-12-04T09:16:48.1007760Z * [new tag] ciflow/trunk/169474 -> ciflow/trunk/169474 2025-12-04T09:16:48.1009203Z * [new tag] ciflow/trunk/169475 -> ciflow/trunk/169475 2025-12-04T09:16:48.1010544Z * [new tag] ciflow/trunk/169476 -> ciflow/trunk/169476 2025-12-04T09:16:48.1012023Z * [new tag] ciflow/trunk/169487 -> ciflow/trunk/169487 2025-12-04T09:16:48.1013383Z * [new tag] ciflow/trunk/169497 -> ciflow/trunk/169497 2025-12-04T09:16:48.1014726Z * [new tag] ciflow/trunk/169503 -> ciflow/trunk/169503 2025-12-04T09:16:48.1016163Z * [new tag] ciflow/trunk/169505 -> ciflow/trunk/169505 2025-12-04T09:16:48.1017484Z * [new tag] ciflow/trunk/169507 -> ciflow/trunk/169507 2025-12-04T09:16:48.1018810Z * [new tag] ciflow/trunk/169514 -> ciflow/trunk/169514 2025-12-04T09:16:48.1020312Z * [new tag] ciflow/trunk/169517 -> ciflow/trunk/169517 2025-12-04T09:16:48.1021602Z * [new tag] ciflow/trunk/169519 -> ciflow/trunk/169519 2025-12-04T09:16:48.1022902Z * [new tag] ciflow/trunk/169528 -> ciflow/trunk/169528 2025-12-04T09:16:48.1024264Z * [new tag] ciflow/trunk/169541 -> ciflow/trunk/169541 2025-12-04T09:16:48.1025728Z * [new tag] ciflow/trunk/169555 -> ciflow/trunk/169555 2025-12-04T09:16:48.1027624Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-12-04T09:16:48.1029240Z * [new tag] ciflow/vllm/165270 -> ciflow/vllm/165270 2025-12-04T09:16:48.1030487Z * [new tag] ciflow/vllm/165274 -> ciflow/vllm/165274 2025-12-04T09:16:48.1031764Z * [new tag] ciflow/vllm/166494 -> ciflow/vllm/166494 2025-12-04T09:16:48.1033069Z * [new tag] ciflow/vllm/169219 -> ciflow/vllm/169219 2025-12-04T09:16:48.1034320Z * [new tag] ciflow/vllm/169220 -> ciflow/vllm/169220 2025-12-04T09:16:48.1035934Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-12-04T09:16:48.1037592Z * [new tag] ciflow/xpu/159718 -> ciflow/xpu/159718 2025-12-04T09:16:48.1038868Z * [new tag] ciflow/xpu/161940 -> ciflow/xpu/161940 2025-12-04T09:16:48.1040217Z * [new tag] ciflow/xpu/163251 -> ciflow/xpu/163251 2025-12-04T09:16:48.1041477Z * [new tag] ciflow/xpu/166829 -> ciflow/xpu/166829 2025-12-04T09:16:48.1042715Z * [new tag] ciflow/xpu/166843 -> ciflow/xpu/166843 2025-12-04T09:16:48.1044037Z * [new tag] ciflow/xpu/167972 -> ciflow/xpu/167972 2025-12-04T09:16:48.1045223Z * [new tag] ciflow/xpu/167981 -> ciflow/xpu/167981 2025-12-04T09:16:48.1046480Z * [new tag] ciflow/xpu/168213 -> ciflow/xpu/168213 2025-12-04T09:16:48.1047837Z * [new tag] ciflow/xpu/168262 -> ciflow/xpu/168262 2025-12-04T09:16:48.1049153Z * [new tag] ciflow/xpu/168328 -> ciflow/xpu/168328 2025-12-04T09:16:48.1050755Z * [new tag] ciflow/xpu/168950 -> ciflow/xpu/168950 2025-12-04T09:16:48.1052607Z * [new tag] ciflow/xpu/169039 -> ciflow/xpu/169039 2025-12-04T09:16:48.1054192Z * [new tag] ciflow/xpu/169200 -> ciflow/xpu/169200 2025-12-04T09:16:48.1055546Z * [new tag] ciflow/xpu/169203 -> ciflow/xpu/169203 2025-12-04T09:16:48.1056938Z * [new tag] ciflow/xpu/169230 -> ciflow/xpu/169230 2025-12-04T09:16:48.1058241Z * [new tag] ciflow/xpu/169231 -> ciflow/xpu/169231 2025-12-04T09:16:48.1059724Z * [new tag] ciflow/xpu/169241 -> ciflow/xpu/169241 2025-12-04T09:16:48.1061097Z * [new tag] ciflow/xpu/169280 -> ciflow/xpu/169280 2025-12-04T09:16:48.1062355Z * [new tag] ciflow/xpu/169296 -> ciflow/xpu/169296 2025-12-04T09:16:48.1063872Z * [new tag] ciflow/xpu/169353 -> ciflow/xpu/169353 2025-12-04T09:16:48.1065767Z * [new tag] ciflow/xpu/169410 -> ciflow/xpu/169410 2025-12-04T09:16:48.1067136Z * [new tag] ciflow/xpu/169442 -> ciflow/xpu/169442 2025-12-04T09:16:48.1068537Z * [new tag] ciflow/xpu/169555 -> ciflow/xpu/169555 2025-12-04T09:16:48.1069870Z * [new tag] cslpull75 -> cslpull75 2025-12-04T09:16:48.1071186Z * [new tag] cslpull76 -> cslpull76 2025-12-04T09:16:48.1072513Z * [new tag] cslpull77 -> cslpull77 2025-12-04T09:16:48.1073944Z * [new tag] cslpull78 -> cslpull78 2025-12-04T09:16:48.1075471Z * [new tag] cslpull79 -> cslpull79 2025-12-04T09:16:48.1077115Z * [new tag] cslpull80 -> cslpull80 2025-12-04T09:16:48.1078562Z * [new tag] cslpull81 -> cslpull81 2025-12-04T09:16:48.1079903Z * [new tag] cslpull82 -> cslpull82 2025-12-04T09:16:48.1081272Z * [new tag] cslpull83 -> cslpull83 2025-12-04T09:16:48.1082669Z * [new tag] cslpull84 -> cslpull84 2025-12-04T09:16:48.1084063Z * [new tag] cslpull85 -> cslpull85 2025-12-04T09:16:48.1085485Z * [new tag] cslpull86 -> cslpull86 2025-12-04T09:16:48.1086857Z * [new tag] cslpull87 -> cslpull87 2025-12-04T09:16:48.1088501Z * [new tag] cslpull88 -> cslpull88 2025-12-04T09:16:48.1090137Z * [new tag] cslpull89 -> cslpull89 2025-12-04T09:16:48.1090953Z * [new tag] cslpull90 -> cslpull90 2025-12-04T09:16:48.1092940Z * [new tag] cslpull91 -> cslpull91 2025-12-04T09:16:48.1094278Z * [new tag] cslpull92 -> cslpull92 2025-12-04T09:16:48.1095797Z * [new tag] flight_5 -> flight_5 2025-12-04T09:16:48.1097798Z * [new tag] flight_5.1 -> flight_5.1 2025-12-04T09:16:48.1099248Z * [new tag] flight_5.2 -> flight_5.2 2025-12-04T09:16:48.1100688Z * [new tag] flight_5.3 -> flight_5.3 2025-12-04T09:16:48.1102183Z * [new tag] forpull1 -> forpull1 2025-12-04T09:16:48.1103923Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-12-04T09:16:48.1105428Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-12-04T09:16:48.1106830Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-12-04T09:16:48.1108345Z * [new tag] nightly-binary -> nightly-binary 2025-12-04T09:16:48.1109792Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-12-04T09:16:48.1111314Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-12-04T09:16:48.1113160Z * [new tag] trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 -> trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 2025-12-04T09:16:48.1114592Z * [new tag] trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e -> trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e 2025-12-04T09:16:48.1116312Z * [new tag] trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 -> trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 2025-12-04T09:16:48.1117958Z * [new tag] trunk/07dcc0b83db3211653a38565a24e15acdba75654 -> trunk/07dcc0b83db3211653a38565a24e15acdba75654 2025-12-04T09:16:48.1119407Z * [new tag] trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb -> trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb 2025-12-04T09:16:48.1120854Z * [new tag] trunk/088048f2fea28ff7d450f65c72419ca45780d30b -> trunk/088048f2fea28ff7d450f65c72419ca45780d30b 2025-12-04T09:16:48.1122159Z * [new tag] trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 -> trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 2025-12-04T09:16:48.1123714Z * [new tag] trunk/0b80a4c62b94402844bf221791c096b0035c6d75 -> trunk/0b80a4c62b94402844bf221791c096b0035c6d75 2025-12-04T09:16:48.1125350Z * [new tag] trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 -> trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 2025-12-04T09:16:48.1126926Z * [new tag] trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 -> trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 2025-12-04T09:16:48.1128087Z * [new tag] trunk/135f3753c418a6879b1954904184937b67e61688 -> trunk/135f3753c418a6879b1954904184937b67e61688 2025-12-04T09:16:48.1129827Z * [new tag] trunk/15da21026cb13cd20257dc9e96830db108743c10 -> trunk/15da21026cb13cd20257dc9e96830db108743c10 2025-12-04T09:16:48.1131078Z * [new tag] trunk/166efdad2ac827f30fb02504c6017520257f88ec -> trunk/166efdad2ac827f30fb02504c6017520257f88ec 2025-12-04T09:16:48.1133741Z * [new tag] trunk/174272c15fae553d8488140af931f7d8050a313f -> trunk/174272c15fae553d8488140af931f7d8050a313f 2025-12-04T09:16:48.1134856Z * [new tag] trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 -> trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 2025-12-04T09:16:48.1136564Z * [new tag] trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 -> trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 2025-12-04T09:16:48.1138147Z * [new tag] trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 -> trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 2025-12-04T09:16:48.1139676Z * [new tag] trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 -> trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 2025-12-04T09:16:48.1140782Z * [new tag] trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e -> trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e 2025-12-04T09:16:48.1142444Z * [new tag] trunk/1c87554d74140eaee964ca8b1832cede67f5f520 -> trunk/1c87554d74140eaee964ca8b1832cede67f5f520 2025-12-04T09:16:48.1143981Z * [new tag] trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 -> trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 2025-12-04T09:16:48.1145463Z * [new tag] trunk/1cee47d6ce0a02227185b566593f002dd639ca0c -> trunk/1cee47d6ce0a02227185b566593f002dd639ca0c 2025-12-04T09:16:48.1146572Z * [new tag] trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d -> trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d 2025-12-04T09:16:48.1148245Z * [new tag] trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 -> trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 2025-12-04T09:16:48.1149955Z * [new tag] trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de -> trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de 2025-12-04T09:16:48.1151254Z * [new tag] trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 -> trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 2025-12-04T09:16:48.1152848Z * [new tag] trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 -> trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 2025-12-04T09:16:48.1154314Z * [new tag] trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f -> trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f 2025-12-04T09:16:48.1155983Z * [new tag] trunk/285779b1621cf9f073a062b0889a642d200308d9 -> trunk/285779b1621cf9f073a062b0889a642d200308d9 2025-12-04T09:16:48.1157021Z * [new tag] trunk/2887faaec6295d081580d09fce161201826c6d87 -> trunk/2887faaec6295d081580d09fce161201826c6d87 2025-12-04T09:16:48.1158769Z * [new tag] trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc -> trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc 2025-12-04T09:16:48.1160247Z * [new tag] trunk/29856679769b3dede478767e2fe6cfb51197cb25 -> trunk/29856679769b3dede478767e2fe6cfb51197cb25 2025-12-04T09:16:48.1161703Z * [new tag] trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 -> trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 2025-12-04T09:16:48.1163178Z * [new tag] trunk/2ac3ef882afb23136adc188975f0a8802fc68adf -> trunk/2ac3ef882afb23136adc188975f0a8802fc68adf 2025-12-04T09:16:48.1164199Z * [new tag] trunk/2bec68e73b64715354af076ad309335f943e36cd -> trunk/2bec68e73b64715354af076ad309335f943e36cd 2025-12-04T09:16:48.1165921Z * [new tag] trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 -> trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 2025-12-04T09:16:48.1167582Z * [new tag] trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 -> trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 2025-12-04T09:16:48.1169069Z * [new tag] trunk/2df6058f116a65722a0e03073402feb242572d35 -> trunk/2df6058f116a65722a0e03073402feb242572d35 2025-12-04T09:16:48.1170312Z * [new tag] trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec -> trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec 2025-12-04T09:16:48.1172228Z * [new tag] trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 -> trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 2025-12-04T09:16:48.1173720Z * [new tag] trunk/305168768a95d69c444df5cd334bb774edfe06f1 -> trunk/305168768a95d69c444df5cd334bb774edfe06f1 2025-12-04T09:16:48.1175130Z * [new tag] trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 -> trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 2025-12-04T09:16:48.1176591Z * [new tag] trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 -> trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 2025-12-04T09:16:48.1178014Z * [new tag] trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 -> trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 2025-12-04T09:16:48.1179451Z * [new tag] trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf -> trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf 2025-12-04T09:16:48.1180928Z * [new tag] trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee -> trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee 2025-12-04T09:16:48.1182444Z * [new tag] trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 -> trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 2025-12-04T09:16:48.1183463Z * [new tag] trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 -> trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 2025-12-04T09:16:48.1185203Z * [new tag] trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae -> trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae 2025-12-04T09:16:48.1186499Z * [new tag] trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f -> trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f 2025-12-04T09:16:48.1188098Z * [new tag] trunk/42e9005cda22da3f1c559c3649218cebd671027c -> trunk/42e9005cda22da3f1c559c3649218cebd671027c 2025-12-04T09:16:48.1189585Z * [new tag] trunk/43b94713bbf340d3c124fde02d0f73add4021247 -> trunk/43b94713bbf340d3c124fde02d0f73add4021247 2025-12-04T09:16:48.1190883Z * [new tag] trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c -> trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c 2025-12-04T09:16:48.1192927Z * [new tag] trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a -> trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a 2025-12-04T09:16:48.1194202Z * [new tag] trunk/45d310ad84854dff730c0b12e577d7998d978686 -> trunk/45d310ad84854dff730c0b12e577d7998d978686 2025-12-04T09:16:48.1196024Z * [new tag] trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 -> trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 2025-12-04T09:16:48.1197306Z * [new tag] trunk/481e5ab336275bd3acd5fa8a611b05b4469012af -> trunk/481e5ab336275bd3acd5fa8a611b05b4469012af 2025-12-04T09:16:48.1198949Z * [new tag] trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 -> trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 2025-12-04T09:16:48.1200482Z * [new tag] trunk/49a04d26088acc17d948ddd66920f3e16371e873 -> trunk/49a04d26088acc17d948ddd66920f3e16371e873 2025-12-04T09:16:48.1201938Z * [new tag] trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 -> trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 2025-12-04T09:16:48.1202996Z * [new tag] trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f -> trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f 2025-12-04T09:16:48.1204874Z * [new tag] trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa -> trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa 2025-12-04T09:16:48.1206493Z * [new tag] trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c -> trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c 2025-12-04T09:16:48.1208415Z * [new tag] trunk/4fefb8e7e942386ffac764a41b232241f82bea3a -> trunk/4fefb8e7e942386ffac764a41b232241f82bea3a 2025-12-04T09:16:48.1209931Z * [new tag] trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d -> trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d 2025-12-04T09:16:48.1211168Z * [new tag] trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 -> trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 2025-12-04T09:16:48.1212849Z * [new tag] trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 -> trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 2025-12-04T09:16:48.1214364Z * [new tag] trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a -> trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a 2025-12-04T09:16:48.1215891Z * [new tag] trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 -> trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 2025-12-04T09:16:48.1217362Z * [new tag] trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 -> trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 2025-12-04T09:16:48.1218865Z * [new tag] trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 -> trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 2025-12-04T09:16:48.1220394Z * [new tag] trunk/5634469fda9e5d98869c82c7d03bb08914245f96 -> trunk/5634469fda9e5d98869c82c7d03bb08914245f96 2025-12-04T09:16:48.1221458Z * [new tag] trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc -> trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc 2025-12-04T09:16:48.1223240Z * [new tag] trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 -> trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 2025-12-04T09:16:48.1224707Z * [new tag] trunk/597930f6b568852356ca9795dac76f9e4653adbd -> trunk/597930f6b568852356ca9795dac76f9e4653adbd 2025-12-04T09:16:48.1225735Z * [new tag] trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 -> trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 2025-12-04T09:16:48.1227588Z * [new tag] trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 -> trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 2025-12-04T09:16:48.1229543Z * [new tag] trunk/5a607febc04c3a2b5824c75f3f60307867439a2c -> trunk/5a607febc04c3a2b5824c75f3f60307867439a2c 2025-12-04T09:16:48.1231021Z * [new tag] trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b -> trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b 2025-12-04T09:16:48.1232077Z * [new tag] trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c -> trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c 2025-12-04T09:16:48.1233769Z * [new tag] trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 -> trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 2025-12-04T09:16:48.1235473Z * [new tag] trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 -> trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 2025-12-04T09:16:48.1237130Z * [new tag] trunk/61be54a31dc09b59d99b62176fb935aee0b924ef -> trunk/61be54a31dc09b59d99b62176fb935aee0b924ef 2025-12-04T09:16:48.1238673Z * [new tag] trunk/62d3ccd71484ed6a760d909b41487101bbc65719 -> trunk/62d3ccd71484ed6a760d909b41487101bbc65719 2025-12-04T09:16:48.1240237Z * [new tag] trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b -> trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b 2025-12-04T09:16:48.1241690Z * [new tag] trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a -> trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a 2025-12-04T09:16:48.1243184Z * [new tag] trunk/66004b993744b4106bf8afaba71f3c228a804206 -> trunk/66004b993744b4106bf8afaba71f3c228a804206 2025-12-04T09:16:48.1244646Z * [new tag] trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 -> trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 2025-12-04T09:16:48.1246279Z * [new tag] trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 -> trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 2025-12-04T09:16:48.1247935Z * [new tag] trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d -> trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d 2025-12-04T09:16:48.1249049Z * [new tag] trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b -> trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b 2025-12-04T09:16:48.1250774Z * [new tag] trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 -> trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 2025-12-04T09:16:48.1252324Z * [new tag] trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 -> trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 2025-12-04T09:16:48.1253782Z * [new tag] trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec -> trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec 2025-12-04T09:16:48.1255239Z * [new tag] trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 -> trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 2025-12-04T09:16:48.1256749Z * [new tag] trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d -> trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d 2025-12-04T09:16:48.1258258Z * [new tag] trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a -> trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a 2025-12-04T09:16:48.1259873Z * [new tag] trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e -> trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e 2025-12-04T09:16:48.1261331Z * [new tag] trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 -> trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 2025-12-04T09:16:48.1262788Z * [new tag] trunk/70d797a5fc109b20a517646fcaa819477cd0d485 -> trunk/70d797a5fc109b20a517646fcaa819477cd0d485 2025-12-04T09:16:48.1264225Z * [new tag] trunk/7348cb355ff0a6f79cd4871215aea72185748734 -> trunk/7348cb355ff0a6f79cd4871215aea72185748734 2025-12-04T09:16:48.1265783Z * [new tag] trunk/74fe26a1ebe32931783569f2e762e3c2c974901f -> trunk/74fe26a1ebe32931783569f2e762e3c2c974901f 2025-12-04T09:16:48.1267262Z * [new tag] trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 -> trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 2025-12-04T09:16:48.1268333Z * [new tag] trunk/7716da9fb23f27a65b41f9f016a2afadf281c18f -> trunk/7716da9fb23f27a65b41f9f016a2afadf281c18f 2025-12-04T09:16:48.1270035Z * [new tag] trunk/7741edd4ed665f3988052e260863efb508d61a03 -> trunk/7741edd4ed665f3988052e260863efb508d61a03 2025-12-04T09:16:48.1271620Z * [new tag] trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 -> trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 2025-12-04T09:16:48.1273338Z * [new tag] trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 -> trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 2025-12-04T09:16:48.1274627Z * [new tag] trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 -> trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 2025-12-04T09:16:48.1276158Z * [new tag] trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca -> trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca 2025-12-04T09:16:48.1277641Z * [new tag] trunk/7b7af390ea8541c611d1ce2018a6934188fc197b -> trunk/7b7af390ea8541c611d1ce2018a6934188fc197b 2025-12-04T09:16:48.1279106Z * [new tag] trunk/7ba4680f3755a560af81aa0f688791e367aa3609 -> trunk/7ba4680f3755a560af81aa0f688791e367aa3609 2025-12-04T09:16:48.1280729Z * [new tag] trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b -> trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b 2025-12-04T09:16:48.1281733Z * [new tag] trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T09:16:48.1283366Z * [new tag] trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 -> trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 2025-12-04T09:16:48.1284887Z * [new tag] trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed -> trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed 2025-12-04T09:16:48.1286426Z * [new tag] trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 -> trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 2025-12-04T09:16:48.1287792Z * [new tag] trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e -> trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e 2025-12-04T09:16:48.1289780Z * [new tag] trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead -> trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead 2025-12-04T09:16:48.1291003Z * [new tag] trunk/81af382128efa094d8702e18f2c133760904c718 -> trunk/81af382128efa094d8702e18f2c133760904c718 2025-12-04T09:16:48.1292920Z * [new tag] trunk/84149583d483e9c973c9a0feda70e4f3964947b0 -> trunk/84149583d483e9c973c9a0feda70e4f3964947b0 2025-12-04T09:16:48.1294809Z * [new tag] trunk/85a315917efe82c24306be805c584ec044951c75 -> trunk/85a315917efe82c24306be805c584ec044951c75 2025-12-04T09:16:48.1296299Z * [new tag] trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece -> trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece 2025-12-04T09:16:48.1297359Z * [new tag] trunk/892640e25aeefa8007c5af837214b4502b6b62a6 -> trunk/892640e25aeefa8007c5af837214b4502b6b62a6 2025-12-04T09:16:48.1299240Z * [new tag] trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 -> trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 2025-12-04T09:16:48.1300671Z * [new tag] trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c -> trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c 2025-12-04T09:16:48.1302173Z * [new tag] trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 -> trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 2025-12-04T09:16:48.1303800Z * [new tag] trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 -> trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 2025-12-04T09:16:48.1305281Z * [new tag] trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca -> trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca 2025-12-04T09:16:48.1306791Z * [new tag] trunk/90b27e7e8352cde97d32ddad24740ef819633f38 -> trunk/90b27e7e8352cde97d32ddad24740ef819633f38 2025-12-04T09:16:48.1307882Z * [new tag] trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 -> trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 2025-12-04T09:16:48.1309494Z * [new tag] trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c -> trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c 2025-12-04T09:16:48.1311082Z * [new tag] trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 -> trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 2025-12-04T09:16:48.1312510Z * [new tag] trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 -> trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 2025-12-04T09:16:48.1314068Z * [new tag] trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa -> trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa 2025-12-04T09:16:48.1315555Z * [new tag] trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d -> trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d 2025-12-04T09:16:48.1317073Z * [new tag] trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 -> trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 2025-12-04T09:16:48.1318513Z * [new tag] trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 -> trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 2025-12-04T09:16:48.1320138Z * [new tag] trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d -> trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d 2025-12-04T09:16:48.1321615Z * [new tag] trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a -> trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a 2025-12-04T09:16:48.1323169Z * [new tag] trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 -> trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 2025-12-04T09:16:48.1324767Z * [new tag] trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 -> trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 2025-12-04T09:16:48.1326251Z * [new tag] trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa -> trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa 2025-12-04T09:16:48.1328087Z * [new tag] trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d -> trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d 2025-12-04T09:16:48.1329563Z * [new tag] trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c -> trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c 2025-12-04T09:16:48.1330975Z * [new tag] trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 -> trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 2025-12-04T09:16:48.1332572Z * [new tag] trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c -> trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c 2025-12-04T09:16:48.1333637Z * [new tag] trunk/a7dc6dab9ad911259d4801c502907e531594db45 -> trunk/a7dc6dab9ad911259d4801c502907e531594db45 2025-12-04T09:16:48.1335633Z * [new tag] trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 -> trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 2025-12-04T09:16:48.1339759Z * [new tag] trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e -> trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e 2025-12-04T09:16:48.1341275Z * [new tag] trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e -> trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e 2025-12-04T09:16:48.1342401Z * [new tag] trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e -> trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e 2025-12-04T09:16:48.1344021Z * [new tag] trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 -> trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 2025-12-04T09:16:48.1345719Z * [new tag] trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 -> trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 2025-12-04T09:16:48.1347323Z * [new tag] trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 -> trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 2025-12-04T09:16:48.1348811Z * [new tag] trunk/b39813b4a04931682b0491adba2138d01d716d99 -> trunk/b39813b4a04931682b0491adba2138d01d716d99 2025-12-04T09:16:48.1350367Z * [new tag] trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 -> trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 2025-12-04T09:16:48.1351966Z * [new tag] trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 -> trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 2025-12-04T09:16:48.1353484Z * [new tag] trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a -> trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a 2025-12-04T09:16:48.1355024Z * [new tag] trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 -> trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 2025-12-04T09:16:48.1356571Z * [new tag] trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 -> trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 2025-12-04T09:16:48.1358188Z * [new tag] trunk/b7d60685f8cbc939b68a20871e90db67e729329b -> trunk/b7d60685f8cbc939b68a20871e90db67e729329b 2025-12-04T09:16:48.1359735Z * [new tag] trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e -> trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e 2025-12-04T09:16:48.1361324Z * [new tag] trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf -> trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf 2025-12-04T09:16:48.1362760Z * [new tag] trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 -> trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 2025-12-04T09:16:48.1364548Z * [new tag] trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f -> trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f 2025-12-04T09:16:48.1366077Z * [new tag] trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f -> trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f 2025-12-04T09:16:48.1367620Z * [new tag] trunk/bb3034198b459401fabeab254e1b99f0115046e2 -> trunk/bb3034198b459401fabeab254e1b99f0115046e2 2025-12-04T09:16:48.1369310Z * [new tag] trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 -> trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 2025-12-04T09:16:48.1371021Z * [new tag] trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 -> trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 2025-12-04T09:16:48.1372344Z * [new tag] trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 -> trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 2025-12-04T09:16:48.1373967Z * [new tag] trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 -> trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 2025-12-04T09:16:48.1375831Z * [new tag] trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 -> trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 2025-12-04T09:16:48.1377294Z * [new tag] trunk/c0660bcee27e7d7731634e274576a7081882bede -> trunk/c0660bcee27e7d7731634e274576a7081882bede 2025-12-04T09:16:48.1378912Z * [new tag] trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac -> trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac 2025-12-04T09:16:48.1380404Z * [new tag] trunk/c55b1e8f61d041ee436d697449eb028931d574fb -> trunk/c55b1e8f61d041ee436d697449eb028931d574fb 2025-12-04T09:16:48.1381699Z * [new tag] trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 -> trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 2025-12-04T09:16:48.1383509Z * [new tag] trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 -> trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 2025-12-04T09:16:48.1385156Z * [new tag] trunk/cc0853af42122f8185321f542616f4474e717f09 -> trunk/cc0853af42122f8185321f542616f4474e717f09 2025-12-04T09:16:48.1386449Z * [new tag] trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 -> trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 2025-12-04T09:16:48.1388141Z * [new tag] trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a -> trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a 2025-12-04T09:16:48.1389782Z * [new tag] trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace -> trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace 2025-12-04T09:16:48.1391680Z * [new tag] trunk/d16447dacaf2420ea175f0c275c75da951f57d39 -> trunk/d16447dacaf2420ea175f0c275c75da951f57d39 2025-12-04T09:16:48.1393228Z * [new tag] trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 -> trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 2025-12-04T09:16:48.1394782Z * [new tag] trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 -> trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 2025-12-04T09:16:48.1396467Z * [new tag] trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf -> trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf 2025-12-04T09:16:48.1397978Z * [new tag] trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 -> trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 2025-12-04T09:16:48.1399504Z * [new tag] trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d -> trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d 2025-12-04T09:16:48.1400984Z * [new tag] trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 -> trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 2025-12-04T09:16:48.1402492Z * [new tag] trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 -> trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 2025-12-04T09:16:48.1404012Z * [new tag] trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e -> trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e 2025-12-04T09:16:48.1405544Z * [new tag] trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a -> trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a 2025-12-04T09:16:48.1407176Z * [new tag] trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b -> trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b 2025-12-04T09:16:48.1408869Z * [new tag] trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec -> trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec 2025-12-04T09:16:48.1410428Z * [new tag] trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf -> trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf 2025-12-04T09:16:48.1411989Z * [new tag] trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd -> trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd 2025-12-04T09:16:48.1413220Z * [new tag] trunk/dd18a75336a4fbd7497955cc5665904724fce889 -> trunk/dd18a75336a4fbd7497955cc5665904724fce889 2025-12-04T09:16:48.1414952Z * [new tag] trunk/ded9bcd61a059bf723e6e84689552962b480ea77 -> trunk/ded9bcd61a059bf723e6e84689552962b480ea77 2025-12-04T09:16:48.1416772Z * [new tag] trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c -> trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c 2025-12-04T09:16:48.1418422Z * [new tag] trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b -> trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b 2025-12-04T09:16:48.1419701Z * [new tag] trunk/e3f24fd73ad74c6e7176687986436956c7c18235 -> trunk/e3f24fd73ad74c6e7176687986436956c7c18235 2025-12-04T09:16:48.1421394Z * [new tag] trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e -> trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e 2025-12-04T09:16:48.1422986Z * [new tag] trunk/ea7035f462a0d2830865ee86c832bd101e1427fc -> trunk/ea7035f462a0d2830865ee86c832bd101e1427fc 2025-12-04T09:16:48.1424576Z * [new tag] trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 -> trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 2025-12-04T09:16:48.1426089Z * [new tag] trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf -> trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf 2025-12-04T09:16:48.1427586Z * [new tag] trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e -> trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e 2025-12-04T09:16:48.1429138Z * [new tag] trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e -> trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e 2025-12-04T09:16:48.1431053Z * [new tag] trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 -> trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 2025-12-04T09:16:48.1432565Z * [new tag] trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 -> trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 2025-12-04T09:16:48.1434136Z * [new tag] trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 -> trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 2025-12-04T09:16:48.1435768Z * [new tag] trunk/f1076f5510920044912247b1abb8760cb820f598 -> trunk/f1076f5510920044912247b1abb8760cb820f598 2025-12-04T09:16:48.1437544Z * [new tag] trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 -> trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 2025-12-04T09:16:48.1439122Z * [new tag] trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 -> trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 2025-12-04T09:16:48.1440636Z * [new tag] trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 -> trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 2025-12-04T09:16:48.1442077Z * [new tag] trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 -> trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 2025-12-04T09:16:48.1443641Z * [new tag] trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 -> trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 2025-12-04T09:16:48.1445141Z * [new tag] trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 -> trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 2025-12-04T09:16:48.1446624Z * [new tag] trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 -> trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 2025-12-04T09:16:48.1448328Z * [new tag] trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b -> trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b 2025-12-04T09:16:48.1449900Z * [new tag] trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 -> trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 2025-12-04T09:16:48.1451865Z * [new tag] trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 -> trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 2025-12-04T09:16:48.1453436Z * [new tag] trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 -> trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 2025-12-04T09:16:48.1455237Z * [new tag] trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 -> trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:16:48.1456027Z * [new tag] v0.1.1 -> v0.1.1 2025-12-04T09:16:48.1457629Z * [new tag] v0.1.10 -> v0.1.10 2025-12-04T09:16:48.1459100Z * [new tag] v0.1.11 -> v0.1.11 2025-12-04T09:16:48.1460563Z * [new tag] v0.1.12 -> v0.1.12 2025-12-04T09:16:48.1461971Z * [new tag] v0.1.2 -> v0.1.2 2025-12-04T09:16:48.1463304Z * [new tag] v0.1.3 -> v0.1.3 2025-12-04T09:16:48.1464631Z * [new tag] v0.1.4 -> v0.1.4 2025-12-04T09:16:48.1466056Z * [new tag] v0.1.5 -> v0.1.5 2025-12-04T09:16:48.1467478Z * [new tag] v0.1.6 -> v0.1.6 2025-12-04T09:16:48.1468814Z * [new tag] v0.1.7 -> v0.1.7 2025-12-04T09:16:48.1470147Z * [new tag] v0.1.8 -> v0.1.8 2025-12-04T09:16:48.1471485Z * [new tag] v0.1.9 -> v0.1.9 2025-12-04T09:16:48.1472956Z * [new tag] v0.2.0 -> v0.2.0 2025-12-04T09:16:48.1474386Z * [new tag] v0.3.0 -> v0.3.0 2025-12-04T09:16:48.1475870Z * [new tag] v0.3.1 -> v0.3.1 2025-12-04T09:16:48.1477275Z * [new tag] v0.4.0 -> v0.4.0 2025-12-04T09:16:48.1478718Z * [new tag] v0.4.1 -> v0.4.1 2025-12-04T09:16:48.1480159Z * [new tag] v1.0.0 -> v1.0.0 2025-12-04T09:16:48.1481635Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-12-04T09:16:48.1483055Z * [new tag] v1.0.1 -> v1.0.1 2025-12-04T09:16:48.1484474Z * [new tag] v1.0rc0 -> v1.0rc0 2025-12-04T09:16:48.1485672Z * [new tag] v1.0rc1 -> v1.0rc1 2025-12-04T09:16:48.1487117Z * [new tag] v1.1.0 -> v1.1.0 2025-12-04T09:16:48.1488666Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-12-04T09:16:48.1490273Z * [new tag] v1.10.0 -> v1.10.0 2025-12-04T09:16:48.1491747Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-12-04T09:16:48.1493288Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-12-04T09:16:48.1494500Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-12-04T09:16:48.1495957Z * [new tag] v1.10.1 -> v1.10.1 2025-12-04T09:16:48.1497166Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-12-04T09:16:48.1498391Z * [new tag] v1.10.2 -> v1.10.2 2025-12-04T09:16:48.1499632Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-12-04T09:16:48.1501062Z * [new tag] v1.11.0 -> v1.11.0 2025-12-04T09:16:48.1502618Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-12-04T09:16:48.1504131Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-12-04T09:16:48.1505743Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-12-04T09:16:48.1507234Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-12-04T09:16:48.1508665Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-12-04T09:16:48.1509910Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-12-04T09:16:48.1510986Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-12-04T09:16:48.1512772Z * [new tag] v1.12.0 -> v1.12.0 2025-12-04T09:16:48.1514031Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-12-04T09:16:48.1515524Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-12-04T09:16:48.1517036Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-12-04T09:16:48.1518505Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-12-04T09:16:48.1519931Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-12-04T09:16:48.1521445Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-12-04T09:16:48.1522663Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-12-04T09:16:48.1523867Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-12-04T09:16:48.1525090Z * [new tag] v1.12.1 -> v1.12.1 2025-12-04T09:16:48.1526630Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-12-04T09:16:48.1528149Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-12-04T09:16:48.1529666Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-12-04T09:16:48.1531164Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-12-04T09:16:48.1532399Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-12-04T09:16:48.1533912Z * [new tag] v1.13.0 -> v1.13.0 2025-12-04T09:16:48.1535434Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-12-04T09:16:48.1537427Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-12-04T09:16:48.1539001Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-12-04T09:16:48.1540556Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-12-04T09:16:48.1541800Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-12-04T09:16:48.1542897Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-12-04T09:16:48.1544513Z * [new tag] v1.13.1 -> v1.13.1 2025-12-04T09:16:48.1545726Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-12-04T09:16:48.1547137Z * [new tag] v1.2.0 -> v1.2.0 2025-12-04T09:16:48.1548618Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-12-04T09:16:48.1550053Z * [new tag] v1.3.0 -> v1.3.0 2025-12-04T09:16:48.1551500Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-12-04T09:16:48.1552705Z * [new tag] v1.3.1 -> v1.3.1 2025-12-04T09:16:48.1554104Z * [new tag] v1.4.0 -> v1.4.0 2025-12-04T09:16:48.1555498Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-12-04T09:16:48.1556713Z * [new tag] v1.4.1 -> v1.4.1 2025-12-04T09:16:48.1558220Z * [new tag] v1.5.0 -> v1.5.0 2025-12-04T09:16:48.1559759Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-12-04T09:16:48.1561238Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-12-04T09:16:48.1562794Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-12-04T09:16:48.1564185Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-12-04T09:16:48.1565410Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-12-04T09:16:48.1566933Z * [new tag] v1.5.1 -> v1.5.1 2025-12-04T09:16:48.1568240Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-12-04T09:16:48.1569611Z * [new tag] v1.6.0 -> v1.6.0 2025-12-04T09:16:48.1570925Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-12-04T09:16:48.1572694Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-12-04T09:16:48.1573978Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-12-04T09:16:48.1575361Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-12-04T09:16:48.1576816Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-12-04T09:16:48.1578306Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-12-04T09:16:48.1579672Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-12-04T09:16:48.1580997Z * [new tag] v1.7.0 -> v1.7.0 2025-12-04T09:16:48.1582444Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-12-04T09:16:48.1584091Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-12-04T09:16:48.1585587Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-12-04T09:16:48.1586810Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-12-04T09:16:48.1588256Z * [new tag] v1.7.1 -> v1.7.1 2025-12-04T09:16:48.1589841Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-12-04T09:16:48.1591356Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-12-04T09:16:48.1592607Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-12-04T09:16:48.1594070Z * [new tag] v1.8.0 -> v1.8.0 2025-12-04T09:16:48.1595290Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-12-04T09:16:48.1596807Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-12-04T09:16:48.1598239Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-12-04T09:16:48.1599639Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-12-04T09:16:48.1600871Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-12-04T09:16:48.1602114Z * [new tag] v1.8.1 -> v1.8.1 2025-12-04T09:16:48.1603611Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-12-04T09:16:48.1604813Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-12-04T09:16:48.1606050Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-12-04T09:16:48.1608012Z * [new tag] v1.8.2 -> v1.8.2 2025-12-04T09:16:48.1609299Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-12-04T09:16:48.1610725Z * [new tag] v1.9.0 -> v1.9.0 2025-12-04T09:16:48.1612229Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-12-04T09:16:48.1613731Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-12-04T09:16:48.1615212Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-12-04T09:16:48.1616466Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-12-04T09:16:48.1617926Z * [new tag] v1.9.1 -> v1.9.1 2025-12-04T09:16:48.1619608Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-12-04T09:16:48.1620820Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-12-04T09:16:48.1622329Z * [new tag] v2.0.0 -> v2.0.0 2025-12-04T09:16:48.1623804Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-12-04T09:16:48.1625276Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-12-04T09:16:48.1626781Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-12-04T09:16:48.1628609Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-12-04T09:16:48.1630225Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-12-04T09:16:48.1631416Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-12-04T09:16:48.1632921Z * [new tag] v2.0.1 -> v2.0.1 2025-12-04T09:16:48.1634458Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-12-04T09:16:48.1635607Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-12-04T09:16:48.1637471Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-12-04T09:16:48.1638718Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-12-04T09:16:48.1640701Z * [new tag] v2.1.0 -> v2.1.0 2025-12-04T09:16:48.1642157Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-12-04T09:16:48.1643665Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-12-04T09:16:48.1645179Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-12-04T09:16:48.1646682Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-12-04T09:16:48.1648288Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-12-04T09:16:48.1649515Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-12-04T09:16:48.1651060Z * [new tag] v2.1.1 -> v2.1.1 2025-12-04T09:16:48.1652630Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-12-04T09:16:48.1654093Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-12-04T09:16:48.1655681Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-12-04T09:16:48.1657121Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-12-04T09:16:48.1658600Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-12-04T09:16:48.1659851Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-12-04T09:16:48.1661215Z * [new tag] v2.1.2 -> v2.1.2 2025-12-04T09:16:48.1662733Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-12-04T09:16:48.1664246Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-12-04T09:16:48.1665490Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-12-04T09:16:48.1667025Z * [new tag] v2.2.0 -> v2.2.0 2025-12-04T09:16:48.1668428Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-12-04T09:16:48.1669904Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-12-04T09:16:48.1671276Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-12-04T09:16:48.1672686Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-12-04T09:16:48.1674096Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-12-04T09:16:48.1675511Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-12-04T09:16:48.1676870Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-12-04T09:16:48.1678106Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-12-04T09:16:48.1679629Z * [new tag] v2.2.1 -> v2.2.1 2025-12-04T09:16:48.1681507Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-12-04T09:16:48.1682715Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-12-04T09:16:48.1683958Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-12-04T09:16:48.1685159Z * [new tag] v2.2.2 -> v2.2.2 2025-12-04T09:16:48.1686758Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-12-04T09:16:48.1688090Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-12-04T09:16:48.1689411Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-12-04T09:16:48.1691060Z * [new tag] v2.3.0 -> v2.3.0 2025-12-04T09:16:48.1692419Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-12-04T09:16:48.1693949Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-12-04T09:16:48.1695547Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-12-04T09:16:48.1696534Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-12-04T09:16:48.1698222Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-12-04T09:16:48.1699801Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-12-04T09:16:48.1701228Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-12-04T09:16:48.1702654Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-12-04T09:16:48.1703973Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-12-04T09:16:48.1705483Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-12-04T09:16:48.1706898Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-12-04T09:16:48.1708113Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-12-04T09:16:48.1709342Z * [new tag] v2.3.1 -> v2.3.1 2025-12-04T09:16:48.1710895Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-12-04T09:16:48.1712388Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-12-04T09:16:48.1713905Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-12-04T09:16:48.1715330Z * [new tag] v2.4.0 -> v2.4.0 2025-12-04T09:16:48.1716818Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-12-04T09:16:48.1718242Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-12-04T09:16:48.1719641Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-12-04T09:16:48.1721544Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-12-04T09:16:48.1723144Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-12-04T09:16:48.1724618Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-12-04T09:16:48.1726137Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-12-04T09:16:48.1727611Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-12-04T09:16:48.1729201Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-12-04T09:16:48.1730320Z * [new tag] v2.4.1 -> v2.4.1 2025-12-04T09:16:48.1732025Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-12-04T09:16:48.1733450Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-12-04T09:16:48.1734969Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-12-04T09:16:48.1739773Z * [new tag] v2.5.0 -> v2.5.0 2025-12-04T09:16:48.1741029Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-12-04T09:16:48.1742249Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-12-04T09:16:48.1743691Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-12-04T09:16:48.1745101Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-12-04T09:16:48.1746544Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-12-04T09:16:48.1748011Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-12-04T09:16:48.1749520Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-12-04T09:16:48.1751018Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-12-04T09:16:48.1752461Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-12-04T09:16:48.1754083Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-12-04T09:16:48.1754965Z * [new tag] v2.5.1 -> v2.5.1 2025-12-04T09:16:48.1756341Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-12-04T09:16:48.1757601Z * [new tag] v2.6.0 -> v2.6.0 2025-12-04T09:16:48.1759175Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-12-04T09:16:48.1760703Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-12-04T09:16:48.1762189Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-12-04T09:16:48.1763569Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-12-04T09:16:48.1765338Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-12-04T09:16:48.1766899Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-12-04T09:16:48.1768537Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-12-04T09:16:48.1770196Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-12-04T09:16:48.1771693Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-12-04T09:16:48.1773436Z * [new tag] v2.7.0 -> v2.7.0 2025-12-04T09:16:48.1774846Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-12-04T09:16:48.1776070Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-12-04T09:16:48.1777638Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-12-04T09:16:48.1779244Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-12-04T09:16:48.1780758Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-12-04T09:16:48.1782236Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-12-04T09:16:48.1783671Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-12-04T09:16:48.1785158Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-12-04T09:16:48.1786668Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-12-04T09:16:48.1788163Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-12-04T09:16:48.1789407Z * [new tag] v2.7.1 -> v2.7.1 2025-12-04T09:16:48.1790974Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-12-04T09:16:48.1792483Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-12-04T09:16:48.1794079Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-12-04T09:16:48.1795616Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-12-04T09:16:48.1797142Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-12-04T09:16:48.1798437Z * [new tag] v2.8.0 -> v2.8.0 2025-12-04T09:16:48.1799979Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-12-04T09:16:48.1801503Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-12-04T09:16:48.1803020Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-12-04T09:16:48.1804614Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-12-04T09:16:48.1806130Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-12-04T09:16:48.1807707Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-12-04T09:16:48.1809316Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-12-04T09:16:48.1810761Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-12-04T09:16:48.1812282Z * [new tag] v2.9.0 -> v2.9.0 2025-12-04T09:16:48.1814009Z * [new tag] v2.9.0-rc1 -> v2.9.0-rc1 2025-12-04T09:16:48.1815496Z * [new tag] v2.9.0-rc10 -> v2.9.0-rc10 2025-12-04T09:16:48.1816891Z * [new tag] v2.9.0-rc11 -> v2.9.0-rc11 2025-12-04T09:16:48.1818646Z * [new tag] v2.9.0-rc2 -> v2.9.0-rc2 2025-12-04T09:16:48.1820675Z * [new tag] v2.9.0-rc3 -> v2.9.0-rc3 2025-12-04T09:16:48.1822345Z * [new tag] v2.9.0-rc4 -> v2.9.0-rc4 2025-12-04T09:16:48.1823882Z * [new tag] v2.9.0-rc5 -> v2.9.0-rc5 2025-12-04T09:16:48.1825630Z * [new tag] v2.9.0-rc6 -> v2.9.0-rc6 2025-12-04T09:16:48.1827131Z * [new tag] v2.9.0-rc7 -> v2.9.0-rc7 2025-12-04T09:16:48.1828830Z * [new tag] v2.9.0-rc8 -> v2.9.0-rc8 2025-12-04T09:16:48.1830145Z * [new tag] v2.9.0-rc9 -> v2.9.0-rc9 2025-12-04T09:16:48.1831436Z * [new tag] v2.9.1 -> v2.9.1 2025-12-04T09:16:48.1833019Z * [new tag] v2.9.1-rc1 -> v2.9.1-rc1 2025-12-04T09:16:48.1834583Z * [new tag] v2.9.1-rc2 -> v2.9.1-rc2 2025-12-04T09:16:48.1837499Z * [new tag] viable/strict/1759343184 -> viable/strict/1759343184 2025-12-04T09:16:48.1838942Z * [new tag] viable/strict/1759346540 -> viable/strict/1759346540 2025-12-04T09:16:48.1840413Z * [new tag] viable/strict/1759348181 -> viable/strict/1759348181 2025-12-04T09:16:48.1841894Z * [new tag] viable/strict/1759350324 -> viable/strict/1759350324 2025-12-04T09:16:48.1843303Z * [new tag] viable/strict/1759351793 -> viable/strict/1759351793 2025-12-04T09:16:48.1844739Z * [new tag] viable/strict/1759353844 -> viable/strict/1759353844 2025-12-04T09:16:48.1846149Z * [new tag] viable/strict/1759355374 -> viable/strict/1759355374 2025-12-04T09:16:48.1847635Z * [new tag] viable/strict/1759357472 -> viable/strict/1759357472 2025-12-04T09:16:48.1849105Z * [new tag] viable/strict/1759361002 -> viable/strict/1759361002 2025-12-04T09:16:48.1850850Z * [new tag] viable/strict/1759362585 -> viable/strict/1759362585 2025-12-04T09:16:48.1852698Z * [new tag] viable/strict/1759365359 -> viable/strict/1759365359 2025-12-04T09:16:48.1854226Z * [new tag] viable/strict/1759370089 -> viable/strict/1759370089 2025-12-04T09:16:48.1855731Z * [new tag] viable/strict/1759377554 -> viable/strict/1759377554 2025-12-04T09:16:48.1857225Z * [new tag] viable/strict/1759379133 -> viable/strict/1759379133 2025-12-04T09:16:48.1858707Z * [new tag] viable/strict/1759389871 -> viable/strict/1759389871 2025-12-04T09:16:48.1860281Z * [new tag] viable/strict/1759393562 -> viable/strict/1759393562 2025-12-04T09:16:48.1861828Z * [new tag] viable/strict/1759395076 -> viable/strict/1759395076 2025-12-04T09:16:48.1863389Z * [new tag] viable/strict/1759398579 -> viable/strict/1759398579 2025-12-04T09:16:48.1864906Z * [new tag] viable/strict/1759404142 -> viable/strict/1759404142 2025-12-04T09:16:48.1866424Z * [new tag] viable/strict/1759405773 -> viable/strict/1759405773 2025-12-04T09:16:48.1867868Z * [new tag] viable/strict/1759408041 -> viable/strict/1759408041 2025-12-04T09:16:48.1869532Z * [new tag] viable/strict/1759411593 -> viable/strict/1759411593 2025-12-04T09:16:48.1871017Z * [new tag] viable/strict/1759427395 -> viable/strict/1759427395 2025-12-04T09:16:48.1872545Z * [new tag] viable/strict/1759434582 -> viable/strict/1759434582 2025-12-04T09:16:48.1874008Z * [new tag] viable/strict/1759436720 -> viable/strict/1759436720 2025-12-04T09:16:48.1875662Z * [new tag] viable/strict/1759440219 -> viable/strict/1759440219 2025-12-04T09:16:48.1877059Z * [new tag] viable/strict/1759441948 -> viable/strict/1759441948 2025-12-04T09:16:48.1878536Z * [new tag] viable/strict/1759443860 -> viable/strict/1759443860 2025-12-04T09:16:48.1880006Z * [new tag] viable/strict/1759445377 -> viable/strict/1759445377 2025-12-04T09:16:48.1881572Z * [new tag] viable/strict/1759447415 -> viable/strict/1759447415 2025-12-04T09:16:48.1883088Z * [new tag] viable/strict/1759451750 -> viable/strict/1759451750 2025-12-04T09:16:48.1884600Z * [new tag] viable/strict/1759453910 -> viable/strict/1759453910 2025-12-04T09:16:48.1886341Z * [new tag] viable/strict/1759456483 -> viable/strict/1759456483 2025-12-04T09:16:48.1888001Z * [new tag] viable/strict/1759459279 -> viable/strict/1759459279 2025-12-04T09:16:48.1889597Z * [new tag] viable/strict/1759460742 -> viable/strict/1759460742 2025-12-04T09:16:48.1891110Z * [new tag] viable/strict/1759462025 -> viable/strict/1759462025 2025-12-04T09:16:48.1892666Z * [new tag] viable/strict/1759469086 -> viable/strict/1759469086 2025-12-04T09:16:48.1894169Z * [new tag] viable/strict/1759470581 -> viable/strict/1759470581 2025-12-04T09:16:48.1895661Z * [new tag] viable/strict/1759472786 -> viable/strict/1759472786 2025-12-04T09:16:48.1897182Z * [new tag] viable/strict/1759476294 -> viable/strict/1759476294 2025-12-04T09:16:48.1898722Z * [new tag] viable/strict/1759479963 -> viable/strict/1759479963 2025-12-04T09:16:48.1900218Z * [new tag] viable/strict/1759492177 -> viable/strict/1759492177 2025-12-04T09:16:48.1901719Z * [new tag] viable/strict/1759519278 -> viable/strict/1759519278 2025-12-04T09:16:48.1903199Z * [new tag] viable/strict/1759524580 -> viable/strict/1759524580 2025-12-04T09:16:48.1904663Z * [new tag] viable/strict/1759528193 -> viable/strict/1759528193 2025-12-04T09:16:48.1906363Z * [new tag] viable/strict/1759533797 -> viable/strict/1759533797 2025-12-04T09:16:48.1907991Z * [new tag] viable/strict/1759542780 -> viable/strict/1759542780 2025-12-04T09:16:48.1909518Z * [new tag] viable/strict/1759549779 -> viable/strict/1759549779 2025-12-04T09:16:48.1911053Z * [new tag] viable/strict/1759555455 -> viable/strict/1759555455 2025-12-04T09:16:48.1912544Z * [new tag] viable/strict/1759559176 -> viable/strict/1759559176 2025-12-04T09:16:48.1914133Z * [new tag] viable/strict/1759560629 -> viable/strict/1759560629 2025-12-04T09:16:48.1915622Z * [new tag] viable/strict/1759569848 -> viable/strict/1759569848 2025-12-04T09:16:48.1917254Z * [new tag] viable/strict/1759571382 -> viable/strict/1759571382 2025-12-04T09:16:48.1918823Z * [new tag] viable/strict/1759573474 -> viable/strict/1759573474 2025-12-04T09:16:48.1920465Z * [new tag] viable/strict/1759618187 -> viable/strict/1759618187 2025-12-04T09:16:48.1921891Z * [new tag] viable/strict/1759626742 -> viable/strict/1759626742 2025-12-04T09:16:48.1923435Z * [new tag] viable/strict/1759632427 -> viable/strict/1759632427 2025-12-04T09:16:48.1924917Z * [new tag] viable/strict/1759634971 -> viable/strict/1759634971 2025-12-04T09:16:48.1926725Z * [new tag] viable/strict/1759661382 -> viable/strict/1759661382 2025-12-04T09:16:48.1928423Z * [new tag] viable/strict/1759663294 -> viable/strict/1759663294 2025-12-04T09:16:48.1929778Z * [new tag] viable/strict/1759708178 -> viable/strict/1759708178 2025-12-04T09:16:48.1931412Z * [new tag] viable/strict/1759715695 -> viable/strict/1759715695 2025-12-04T09:16:48.1932831Z * [new tag] viable/strict/1759728293 -> viable/strict/1759728293 2025-12-04T09:16:48.1934767Z * [new tag] viable/strict/1759735513 -> viable/strict/1759735513 2025-12-04T09:16:48.1936439Z * [new tag] viable/strict/1759739177 -> viable/strict/1759739177 2025-12-04T09:16:48.1940781Z * [new tag] viable/strict/1759758635 -> viable/strict/1759758635 2025-12-04T09:16:48.1942350Z * [new tag] viable/strict/1759765784 -> viable/strict/1759765784 2025-12-04T09:16:48.1943812Z * [new tag] viable/strict/1759767948 -> viable/strict/1759767948 2025-12-04T09:16:48.1945349Z * [new tag] viable/strict/1759771461 -> viable/strict/1759771461 2025-12-04T09:16:48.1946672Z * [new tag] viable/strict/1759776706 -> viable/strict/1759776706 2025-12-04T09:16:48.1948290Z * [new tag] viable/strict/1759782317 -> viable/strict/1759782317 2025-12-04T09:16:48.1949820Z * [new tag] viable/strict/1759783777 -> viable/strict/1759783777 2025-12-04T09:16:48.1951353Z * [new tag] viable/strict/1759785815 -> viable/strict/1759785815 2025-12-04T09:16:48.1952977Z * [new tag] viable/strict/1759789459 -> viable/strict/1759789459 2025-12-04T09:16:48.1954452Z * [new tag] viable/strict/1759790974 -> viable/strict/1759790974 2025-12-04T09:16:48.1955818Z * [new tag] viable/strict/1759794583 -> viable/strict/1759794583 2025-12-04T09:16:48.1957319Z * [new tag] viable/strict/1759797408 -> viable/strict/1759797408 2025-12-04T09:16:48.1958851Z * [new tag] viable/strict/1759799518 -> viable/strict/1759799518 2025-12-04T09:16:48.1960395Z * [new tag] viable/strict/1759804909 -> viable/strict/1759804909 2025-12-04T09:16:48.1961880Z * [new tag] viable/strict/1759807643 -> viable/strict/1759807643 2025-12-04T09:16:48.1963427Z * [new tag] viable/strict/1759809089 -> viable/strict/1759809089 2025-12-04T09:16:48.1964923Z * [new tag] viable/strict/1759811145 -> viable/strict/1759811145 2025-12-04T09:16:48.1966524Z * [new tag] viable/strict/1759812581 -> viable/strict/1759812581 2025-12-04T09:16:48.1968438Z * [new tag] viable/strict/1759814683 -> viable/strict/1759814683 2025-12-04T09:16:48.1969972Z * [new tag] viable/strict/1759821889 -> viable/strict/1759821889 2025-12-04T09:16:48.1971698Z * [new tag] viable/strict/1759823376 -> viable/strict/1759823376 2025-12-04T09:16:48.1973175Z * [new tag] viable/strict/1759827107 -> viable/strict/1759827107 2025-12-04T09:16:48.1974658Z * [new tag] viable/strict/1759830577 -> viable/strict/1759830577 2025-12-04T09:16:48.1976229Z * [new tag] viable/strict/1759832720 -> viable/strict/1759832720 2025-12-04T09:16:48.1977700Z * [new tag] viable/strict/1759842063 -> viable/strict/1759842063 2025-12-04T09:16:48.1979265Z * [new tag] viable/strict/1759847121 -> viable/strict/1759847121 2025-12-04T09:16:48.1989592Z * [new tag] viable/strict/1759850721 -> viable/strict/1759850721 2025-12-04T09:16:48.1989956Z * [new tag] viable/strict/1759857870 -> viable/strict/1759857870 2025-12-04T09:16:48.1990161Z * [new tag] viable/strict/1759863143 -> viable/strict/1759863143 2025-12-04T09:16:48.1990352Z * [new tag] viable/strict/1759875874 -> viable/strict/1759875874 2025-12-04T09:16:48.1990532Z * [new tag] viable/strict/1759877385 -> viable/strict/1759877385 2025-12-04T09:16:48.1990717Z * [new tag] viable/strict/1759883801 -> viable/strict/1759883801 2025-12-04T09:16:48.1991075Z * [new tag] viable/strict/1759885922 -> viable/strict/1759885922 2025-12-04T09:16:48.1991582Z * [new tag] viable/strict/1759888488 -> viable/strict/1759888488 2025-12-04T09:16:48.1993262Z * [new tag] viable/strict/1759895471 -> viable/strict/1759895471 2025-12-04T09:16:48.1994778Z * [new tag] viable/strict/1759904803 -> viable/strict/1759904803 2025-12-04T09:16:48.1996445Z * [new tag] viable/strict/1759908300 -> viable/strict/1759908300 2025-12-04T09:16:48.1997995Z * [new tag] viable/strict/1759915520 -> viable/strict/1759915520 2025-12-04T09:16:48.1999515Z * [new tag] viable/strict/1759916978 -> viable/strict/1759916978 2025-12-04T09:16:48.2000848Z * [new tag] viable/strict/1759930024 -> viable/strict/1759930024 2025-12-04T09:16:48.2002379Z * [new tag] viable/strict/1759948122 -> viable/strict/1759948122 2025-12-04T09:16:48.2004016Z * [new tag] viable/strict/1759952983 -> viable/strict/1759952983 2025-12-04T09:16:48.2005545Z * [new tag] viable/strict/1759955121 -> viable/strict/1759955121 2025-12-04T09:16:48.2007060Z * [new tag] viable/strict/1759962298 -> viable/strict/1759962298 2025-12-04T09:16:48.2008731Z * [new tag] viable/strict/1759965837 -> viable/strict/1759965837 2025-12-04T09:16:48.2010312Z * [new tag] viable/strict/1759970213 -> viable/strict/1759970213 2025-12-04T09:16:48.2011843Z * [new tag] viable/strict/1759974894 -> viable/strict/1759974894 2025-12-04T09:16:48.2013275Z * [new tag] viable/strict/1759977763 -> viable/strict/1759977763 2025-12-04T09:16:48.2014902Z * [new tag] viable/strict/1759979241 -> viable/strict/1759979241 2025-12-04T09:16:48.2016398Z * [new tag] viable/strict/1759985417 -> viable/strict/1759985417 2025-12-04T09:16:48.2017977Z * [new tag] viable/strict/1759987490 -> viable/strict/1759987490 2025-12-04T09:16:48.2019566Z * [new tag] viable/strict/1759996180 -> viable/strict/1759996180 2025-12-04T09:16:48.2021173Z * [new tag] viable/strict/1760065682 -> viable/strict/1760065682 2025-12-04T09:16:48.2022767Z * [new tag] viable/strict/1760066894 -> viable/strict/1760066894 2025-12-04T09:16:48.2024254Z * [new tag] viable/strict/1760070345 -> viable/strict/1760070345 2025-12-04T09:16:48.2025869Z * [new tag] viable/strict/1760089782 -> viable/strict/1760089782 2025-12-04T09:16:48.2027396Z * [new tag] viable/strict/1760091921 -> viable/strict/1760091921 2025-12-04T09:16:48.2028913Z * [new tag] viable/strict/1760127924 -> viable/strict/1760127924 2025-12-04T09:16:48.2030494Z * [new tag] viable/strict/1760129489 -> viable/strict/1760129489 2025-12-04T09:16:48.2032099Z * [new tag] viable/strict/1760132980 -> viable/strict/1760132980 2025-12-04T09:16:48.2033739Z * [new tag] viable/strict/1760135060 -> viable/strict/1760135060 2025-12-04T09:16:48.2035208Z * [new tag] viable/strict/1760215782 -> viable/strict/1760215782 2025-12-04T09:16:48.2037548Z * [new tag] viable/strict/1760273849 -> viable/strict/1760273849 2025-12-04T09:16:48.2038966Z * [new tag] viable/strict/1760275517 -> viable/strict/1760275517 2025-12-04T09:16:48.2040547Z * [new tag] viable/strict/1760276979 -> viable/strict/1760276979 2025-12-04T09:16:48.2042067Z * [new tag] viable/strict/1760279007 -> viable/strict/1760279007 2025-12-04T09:16:48.2043444Z * [new tag] viable/strict/1760286328 -> viable/strict/1760286328 2025-12-04T09:16:48.2044800Z * [new tag] viable/strict/1760493304 -> viable/strict/1760493304 2025-12-04T09:16:48.2046613Z * [new tag] viable/strict/1760496298 -> viable/strict/1760496298 2025-12-04T09:16:48.2047972Z * [new tag] viable/strict/1760518396 -> viable/strict/1760518396 2025-12-04T09:16:48.2049539Z * [new tag] viable/strict/1760534864 -> viable/strict/1760534864 2025-12-04T09:16:48.2051119Z * [new tag] viable/strict/1760549062 -> viable/strict/1760549062 2025-12-04T09:16:48.2052784Z * [new tag] viable/strict/1760552799 -> viable/strict/1760552799 2025-12-04T09:16:48.2054276Z * [new tag] viable/strict/1760554355 -> viable/strict/1760554355 2025-12-04T09:16:48.2055841Z * [new tag] viable/strict/1760556275 -> viable/strict/1760556275 2025-12-04T09:16:48.2057333Z * [new tag] viable/strict/1760564979 -> viable/strict/1760564979 2025-12-04T09:16:48.2058973Z * [new tag] viable/strict/1760567049 -> viable/strict/1760567049 2025-12-04T09:16:48.2060907Z * [new tag] viable/strict/1760568585 -> viable/strict/1760568585 2025-12-04T09:16:48.2062377Z * [new tag] viable/strict/1760570630 -> viable/strict/1760570630 2025-12-04T09:16:48.2063908Z * [new tag] viable/strict/1760572180 -> viable/strict/1760572180 2025-12-04T09:16:48.2065527Z * [new tag] viable/strict/1760575094 -> viable/strict/1760575094 2025-12-04T09:16:48.2067165Z * [new tag] viable/strict/1760579709 -> viable/strict/1760579709 2025-12-04T09:16:48.2069164Z * [new tag] viable/strict/1760582614 -> viable/strict/1760582614 2025-12-04T09:16:48.2070645Z * [new tag] viable/strict/1760586815 -> viable/strict/1760586815 2025-12-04T09:16:48.2072108Z * [new tag] viable/strict/1760588829 -> viable/strict/1760588829 2025-12-04T09:16:48.2073635Z * [new tag] viable/strict/1760590200 -> viable/strict/1760590200 2025-12-04T09:16:48.2075307Z * [new tag] viable/strict/1760592311 -> viable/strict/1760592311 2025-12-04T09:16:48.2076814Z * [new tag] viable/strict/1760619733 -> viable/strict/1760619733 2025-12-04T09:16:48.2078167Z * [new tag] viable/strict/1760628335 -> viable/strict/1760628335 2025-12-04T09:16:48.2079663Z * [new tag] viable/strict/1760635490 -> viable/strict/1760635490 2025-12-04T09:16:48.2081154Z * [new tag] viable/strict/1760640743 -> viable/strict/1760640743 2025-12-04T09:16:48.2082671Z * [new tag] viable/strict/1760642528 -> viable/strict/1760642528 2025-12-04T09:16:48.2084182Z * [new tag] viable/strict/1760646330 -> viable/strict/1760646330 2025-12-04T09:16:48.2085694Z * [new tag] viable/strict/1760666101 -> viable/strict/1760666101 2025-12-04T09:16:48.2087226Z * [new tag] viable/strict/1760668990 -> viable/strict/1760668990 2025-12-04T09:16:48.2088938Z * [new tag] viable/strict/1760670600 -> viable/strict/1760670600 2025-12-04T09:16:48.2090449Z * [new tag] viable/strict/1760671704 -> viable/strict/1760671704 2025-12-04T09:16:48.2091965Z * [new tag] viable/strict/1760673121 -> viable/strict/1760673121 2025-12-04T09:16:48.2093481Z * [new tag] viable/strict/1760675352 -> viable/strict/1760675352 2025-12-04T09:16:48.2095070Z * [new tag] viable/strict/1760696731 -> viable/strict/1760696731 2025-12-04T09:16:48.2097942Z * [new tag] viable/strict/1760723515 -> viable/strict/1760723515 2025-12-04T09:16:48.2099456Z * [new tag] viable/strict/1760727234 -> viable/strict/1760727234 2025-12-04T09:16:48.2101002Z * [new tag] viable/strict/1760730578 -> viable/strict/1760730578 2025-12-04T09:16:48.2102553Z * [new tag] viable/strict/1760732726 -> viable/strict/1760732726 2025-12-04T09:16:48.2104202Z * [new tag] viable/strict/1760734180 -> viable/strict/1760734180 2025-12-04T09:16:48.2105708Z * [new tag] viable/strict/1760736251 -> viable/strict/1760736251 2025-12-04T09:16:48.2107238Z * [new tag] viable/strict/1760737772 -> viable/strict/1760737772 2025-12-04T09:16:48.2108736Z * [new tag] viable/strict/1760758005 -> viable/strict/1760758005 2025-12-04T09:16:48.2110223Z * [new tag] viable/strict/1760761532 -> viable/strict/1760761532 2025-12-04T09:16:48.2111777Z * [new tag] viable/strict/1760802581 -> viable/strict/1760802581 2025-12-04T09:16:48.2113313Z * [new tag] viable/strict/1760827772 -> viable/strict/1760827772 2025-12-04T09:16:48.2114845Z * [new tag] viable/strict/1760834524 -> viable/strict/1760834524 2025-12-04T09:16:48.2116452Z * [new tag] viable/strict/1760845009 -> viable/strict/1760845009 2025-12-04T09:16:48.2118011Z * [new tag] viable/strict/1760876836 -> viable/strict/1760876836 2025-12-04T09:16:48.2119522Z * [new tag] viable/strict/1760880329 -> viable/strict/1760880329 2025-12-04T09:16:48.2120988Z * [new tag] viable/strict/1760888987 -> viable/strict/1760888987 2025-12-04T09:16:48.2122571Z * [new tag] viable/strict/1760912664 -> viable/strict/1760912664 2025-12-04T09:16:48.2124113Z * [new tag] viable/strict/1760925321 -> viable/strict/1760925321 2025-12-04T09:16:48.2125630Z * [new tag] viable/strict/1760931488 -> viable/strict/1760931488 2025-12-04T09:16:48.2127246Z * [new tag] viable/strict/1760932693 -> viable/strict/1760932693 2025-12-04T09:16:48.2128914Z * [new tag] viable/strict/1761004184 -> viable/strict/1761004184 2025-12-04T09:16:48.2130438Z * [new tag] viable/strict/1761014748 -> viable/strict/1761014748 2025-12-04T09:16:48.2131958Z * [new tag] viable/strict/1761017491 -> viable/strict/1761017491 2025-12-04T09:16:48.2133531Z * [new tag] viable/strict/1761018806 -> viable/strict/1761018806 2025-12-04T09:16:48.2135073Z * [new tag] viable/strict/1761020754 -> viable/strict/1761020754 2025-12-04T09:16:48.2136940Z * [new tag] viable/strict/1761024303 -> viable/strict/1761024303 2025-12-04T09:16:48.2138900Z * [new tag] viable/strict/1761029582 -> viable/strict/1761029582 2025-12-04T09:16:48.2140457Z * [new tag] viable/strict/1761031535 -> viable/strict/1761031535 2025-12-04T09:16:48.2141891Z * [new tag] viable/strict/1761035196 -> viable/strict/1761035196 2025-12-04T09:16:48.2143535Z * [new tag] viable/strict/1761045825 -> viable/strict/1761045825 2025-12-04T09:16:48.2145149Z * [new tag] viable/strict/1761054796 -> viable/strict/1761054796 2025-12-04T09:16:48.2146700Z * [new tag] viable/strict/1761060314 -> viable/strict/1761060314 2025-12-04T09:16:48.2148159Z * [new tag] viable/strict/1761071198 -> viable/strict/1761071198 2025-12-04T09:16:48.2149772Z * [new tag] viable/strict/1761074628 -> viable/strict/1761074628 2025-12-04T09:16:48.2151301Z * [new tag] viable/strict/1761078351 -> viable/strict/1761078351 2025-12-04T09:16:48.2153207Z * [new tag] viable/strict/1761079822 -> viable/strict/1761079822 2025-12-04T09:16:48.2154751Z * [new tag] viable/strict/1761081873 -> viable/strict/1761081873 2025-12-04T09:16:48.2156318Z * [new tag] viable/strict/1761083392 -> viable/strict/1761083392 2025-12-04T09:16:48.2157838Z * [new tag] viable/strict/1761085465 -> viable/strict/1761085465 2025-12-04T09:16:48.2159485Z * [new tag] viable/strict/1761089099 -> viable/strict/1761089099 2025-12-04T09:16:48.2161060Z * [new tag] viable/strict/1761095535 -> viable/strict/1761095535 2025-12-04T09:16:48.2162424Z * [new tag] viable/strict/1761098119 -> viable/strict/1761098119 2025-12-04T09:16:48.2164406Z * [new tag] viable/strict/1761101330 -> viable/strict/1761101330 2025-12-04T09:16:48.2165972Z * [new tag] viable/strict/1761114425 -> viable/strict/1761114425 2025-12-04T09:16:48.2167562Z * [new tag] viable/strict/1761116036 -> viable/strict/1761116036 2025-12-04T09:16:48.2169197Z * [new tag] viable/strict/1761119379 -> viable/strict/1761119379 2025-12-04T09:16:48.2170718Z * [new tag] viable/strict/1761121601 -> viable/strict/1761121601 2025-12-04T09:16:48.2172223Z * [new tag] viable/strict/1761123234 -> viable/strict/1761123234 2025-12-04T09:16:48.2173839Z * [new tag] viable/strict/1761126621 -> viable/strict/1761126621 2025-12-04T09:16:48.2175366Z * [new tag] viable/strict/1761132259 -> viable/strict/1761132259 2025-12-04T09:16:48.2176933Z * [new tag] viable/strict/1761146746 -> viable/strict/1761146746 2025-12-04T09:16:48.2178633Z * [new tag] viable/strict/1761164752 -> viable/strict/1761164752 2025-12-04T09:16:48.2180236Z * [new tag] viable/strict/1761166198 -> viable/strict/1761166198 2025-12-04T09:16:48.2181815Z * [new tag] viable/strict/1761175424 -> viable/strict/1761175424 2025-12-04T09:16:48.2183331Z * [new tag] viable/strict/1761176983 -> viable/strict/1761176983 2025-12-04T09:16:48.2185014Z * [new tag] viable/strict/1761179891 -> viable/strict/1761179891 2025-12-04T09:16:48.2186609Z * [new tag] viable/strict/1761181930 -> viable/strict/1761181930 2025-12-04T09:16:48.2188132Z * [new tag] viable/strict/1761184516 -> viable/strict/1761184516 2025-12-04T09:16:48.2189752Z * [new tag] viable/strict/1761190179 -> viable/strict/1761190179 2025-12-04T09:16:48.2191262Z * [new tag] viable/strict/1761193558 -> viable/strict/1761193558 2025-12-04T09:16:48.2192831Z * [new tag] viable/strict/1761207990 -> viable/strict/1761207990 2025-12-04T09:16:48.2194381Z * [new tag] viable/strict/1761229539 -> viable/strict/1761229539 2025-12-04T09:16:48.2196103Z * [new tag] viable/strict/1761244031 -> viable/strict/1761244031 2025-12-04T09:16:48.2197706Z * [new tag] viable/strict/1761248986 -> viable/strict/1761248986 2025-12-04T09:16:48.2199266Z * [new tag] viable/strict/1761259791 -> viable/strict/1761259791 2025-12-04T09:16:48.2200764Z * [new tag] viable/strict/1761266139 -> viable/strict/1761266139 2025-12-04T09:16:48.2202297Z * [new tag] viable/strict/1761268316 -> viable/strict/1761268316 2025-12-04T09:16:48.2203829Z * [new tag] viable/strict/1761273805 -> viable/strict/1761273805 2025-12-04T09:16:48.2205342Z * [new tag] viable/strict/1761275261 -> viable/strict/1761275261 2025-12-04T09:16:48.2206885Z * [new tag] viable/strict/1761277913 -> viable/strict/1761277913 2025-12-04T09:16:48.2208651Z * [new tag] viable/strict/1761290701 -> viable/strict/1761290701 2025-12-04T09:16:48.2210253Z * [new tag] viable/strict/1761294396 -> viable/strict/1761294396 2025-12-04T09:16:48.2211927Z * [new tag] viable/strict/1761303047 -> viable/strict/1761303047 2025-12-04T09:16:48.2213419Z * [new tag] viable/strict/1761335388 -> viable/strict/1761335388 2025-12-04T09:16:48.2214993Z * [new tag] viable/strict/1761337551 -> viable/strict/1761337551 2025-12-04T09:16:48.2216664Z * [new tag] viable/strict/1761339007 -> viable/strict/1761339007 2025-12-04T09:16:48.2218060Z * [new tag] viable/strict/1761341050 -> viable/strict/1761341050 2025-12-04T09:16:48.2219689Z * [new tag] viable/strict/1761346188 -> viable/strict/1761346188 2025-12-04T09:16:48.2221350Z * [new tag] viable/strict/1761349792 -> viable/strict/1761349792 2025-12-04T09:16:48.2222935Z * [new tag] viable/strict/1761352620 -> viable/strict/1761352620 2025-12-04T09:16:48.2224509Z * [new tag] viable/strict/1761354730 -> viable/strict/1761354730 2025-12-04T09:16:48.2226059Z * [new tag] viable/strict/1761357298 -> viable/strict/1761357298 2025-12-04T09:16:48.2227620Z * [new tag] viable/strict/1761360201 -> viable/strict/1761360201 2025-12-04T09:16:48.2229127Z * [new tag] viable/strict/1761361753 -> viable/strict/1761361753 2025-12-04T09:16:48.2230670Z * [new tag] viable/strict/1761364351 -> viable/strict/1761364351 2025-12-04T09:16:48.2232143Z * [new tag] viable/strict/1761366338 -> viable/strict/1761366338 2025-12-04T09:16:48.2233857Z * [new tag] viable/strict/1761367802 -> viable/strict/1761367802 2025-12-04T09:16:48.2235663Z * [new tag] viable/strict/1761369889 -> viable/strict/1761369889 2025-12-04T09:16:48.2237389Z * [new tag] viable/strict/1761371385 -> viable/strict/1761371385 2025-12-04T09:16:48.2239032Z * [new tag] viable/strict/1761373581 -> viable/strict/1761373581 2025-12-04T09:16:48.2241168Z * [new tag] viable/strict/1761375054 -> viable/strict/1761375054 2025-12-04T09:16:48.2242729Z * [new tag] viable/strict/1761421785 -> viable/strict/1761421785 2025-12-04T09:16:48.2244399Z * [new tag] viable/strict/1761434614 -> viable/strict/1761434614 2025-12-04T09:16:48.2246281Z * [new tag] viable/strict/1761439254 -> viable/strict/1761439254 2025-12-04T09:16:48.2248064Z * [new tag] viable/strict/1761454187 -> viable/strict/1761454187 2025-12-04T09:16:48.2249716Z * [new tag] viable/strict/1761459991 -> viable/strict/1761459991 2025-12-04T09:16:48.2251445Z * [new tag] viable/strict/1761470668 -> viable/strict/1761470668 2025-12-04T09:16:48.2253382Z * [new tag] viable/strict/1761472188 -> viable/strict/1761472188 2025-12-04T09:16:48.2254935Z * [new tag] viable/strict/1761503178 -> viable/strict/1761503178 2025-12-04T09:16:48.2256472Z * [new tag] viable/strict/1761517492 -> viable/strict/1761517492 2025-12-04T09:16:48.2258030Z * [new tag] viable/strict/1761518981 -> viable/strict/1761518981 2025-12-04T09:16:48.2259597Z * [new tag] viable/strict/1761533609 -> viable/strict/1761533609 2025-12-04T09:16:48.2261068Z * [new tag] viable/strict/1761546438 -> viable/strict/1761546438 2025-12-04T09:16:48.2262678Z * [new tag] viable/strict/1761548133 -> viable/strict/1761548133 2025-12-04T09:16:48.2264450Z * [new tag] viable/strict/1761555186 -> viable/strict/1761555186 2025-12-04T09:16:48.2266105Z * [new tag] viable/strict/1761557178 -> viable/strict/1761557178 2025-12-04T09:16:48.2267664Z * [new tag] viable/strict/1761560772 -> viable/strict/1761560772 2025-12-04T09:16:48.2269232Z * [new tag] viable/strict/1761562266 -> viable/strict/1761562266 2025-12-04T09:16:48.2270848Z * [new tag] viable/strict/1761564260 -> viable/strict/1761564260 2025-12-04T09:16:48.2272395Z * [new tag] viable/strict/1761568072 -> viable/strict/1761568072 2025-12-04T09:16:48.2273984Z * [new tag] viable/strict/1761571683 -> viable/strict/1761571683 2025-12-04T09:16:48.2275529Z * [new tag] viable/strict/1761580199 -> viable/strict/1761580199 2025-12-04T09:16:48.2277120Z * [new tag] viable/strict/1761587383 -> viable/strict/1761587383 2025-12-04T09:16:48.2278742Z * [new tag] viable/strict/1761591165 -> viable/strict/1761591165 2025-12-04T09:16:48.2280438Z * [new tag] viable/strict/1761594575 -> viable/strict/1761594575 2025-12-04T09:16:48.2282023Z * [new tag] viable/strict/1761596710 -> viable/strict/1761596710 2025-12-04T09:16:48.2283553Z * [new tag] viable/strict/1761598189 -> viable/strict/1761598189 2025-12-04T09:16:48.2285157Z * [new tag] viable/strict/1761600254 -> viable/strict/1761600254 2025-12-04T09:16:48.2286697Z * [new tag] viable/strict/1761603879 -> viable/strict/1761603879 2025-12-04T09:16:48.2288391Z * [new tag] viable/strict/1761605429 -> viable/strict/1761605429 2025-12-04T09:16:48.2290042Z * [new tag] viable/strict/1761607468 -> viable/strict/1761607468 2025-12-04T09:16:48.2291748Z * [new tag] viable/strict/1761608983 -> viable/strict/1761608983 2025-12-04T09:16:48.2293305Z * [new tag] viable/strict/1761611846 -> viable/strict/1761611846 2025-12-04T09:16:48.2294848Z * [new tag] viable/strict/1761613922 -> viable/strict/1761613922 2025-12-04T09:16:48.2296230Z * [new tag] viable/strict/1761616504 -> viable/strict/1761616504 2025-12-04T09:16:48.2297644Z * [new tag] viable/strict/1761619599 -> viable/strict/1761619599 2025-12-04T09:16:48.2299174Z * [new tag] viable/strict/1761686693 -> viable/strict/1761686693 2025-12-04T09:16:48.2300764Z * [new tag] viable/strict/1761688179 -> viable/strict/1761688179 2025-12-04T09:16:48.2302250Z * [new tag] viable/strict/1761691973 -> viable/strict/1761691973 2025-12-04T09:16:48.2304016Z * [new tag] viable/strict/1761693884 -> viable/strict/1761693884 2025-12-04T09:16:48.2305541Z * [new tag] viable/strict/1761695389 -> viable/strict/1761695389 2025-12-04T09:16:48.2307096Z * [new tag] viable/strict/1761698408 -> viable/strict/1761698408 2025-12-04T09:16:48.2308593Z * [new tag] viable/strict/1761702931 -> viable/strict/1761702931 2025-12-04T09:16:48.2310254Z * [new tag] viable/strict/1761706307 -> viable/strict/1761706307 2025-12-04T09:16:48.2311840Z * [new tag] viable/strict/1761709065 -> viable/strict/1761709065 2025-12-04T09:16:48.2313472Z * [new tag] viable/strict/1761710285 -> viable/strict/1761710285 2025-12-04T09:16:48.2315107Z * [new tag] viable/strict/1761711983 -> viable/strict/1761711983 2025-12-04T09:16:48.2316752Z * [new tag] viable/strict/1761713514 -> viable/strict/1761713514 2025-12-04T09:16:48.2318394Z * [new tag] viable/strict/1761715523 -> viable/strict/1761715523 2025-12-04T09:16:48.2320043Z * [new tag] viable/strict/1761727973 -> viable/strict/1761727973 2025-12-04T09:16:48.2321626Z * [new tag] viable/strict/1761751558 -> viable/strict/1761751558 2025-12-04T09:16:48.2323257Z * [new tag] viable/strict/1761755187 -> viable/strict/1761755187 2025-12-04T09:16:48.2324841Z * [new tag] viable/strict/1761756826 -> viable/strict/1761756826 2025-12-04T09:16:48.2326483Z * [new tag] viable/strict/1761769551 -> viable/strict/1761769551 2025-12-04T09:16:48.2328347Z * [new tag] viable/strict/1761771032 -> viable/strict/1761771032 2025-12-04T09:16:48.2330016Z * [new tag] viable/strict/1761773101 -> viable/strict/1761773101 2025-12-04T09:16:48.2331600Z * [new tag] viable/strict/1761781792 -> viable/strict/1761781792 2025-12-04T09:16:48.2333329Z * [new tag] viable/strict/1761784788 -> viable/strict/1761784788 2025-12-04T09:16:48.2334857Z * [new tag] viable/strict/1761786740 -> viable/strict/1761786740 2025-12-04T09:16:48.2339399Z * [new tag] viable/strict/1761789332 -> viable/strict/1761789332 2025-12-04T09:16:48.2341486Z * [new tag] viable/strict/1761792569 -> viable/strict/1761792569 2025-12-04T09:16:48.2343087Z * [new tag] viable/strict/1761795289 -> viable/strict/1761795289 2025-12-04T09:16:48.2344703Z * [new tag] viable/strict/1761798345 -> viable/strict/1761798345 2025-12-04T09:16:48.2346318Z * [new tag] viable/strict/1761799827 -> viable/strict/1761799827 2025-12-04T09:16:48.2348494Z * [new tag] viable/strict/1761805604 -> viable/strict/1761805604 2025-12-04T09:16:48.2350101Z * [new tag] viable/strict/1761807202 -> viable/strict/1761807202 2025-12-04T09:16:48.2351705Z * [new tag] viable/strict/1761809094 -> viable/strict/1761809094 2025-12-04T09:16:48.2353314Z * [new tag] viable/strict/1761810576 -> viable/strict/1761810576 2025-12-04T09:16:48.2354923Z * [new tag] viable/strict/1761812771 -> viable/strict/1761812771 2025-12-04T09:16:48.2356517Z * [new tag] viable/strict/1761814363 -> viable/strict/1761814363 2025-12-04T09:16:48.2358060Z * [new tag] viable/strict/1761857410 -> viable/strict/1761857410 2025-12-04T09:16:48.2359675Z * [new tag] viable/strict/1761860985 -> viable/strict/1761860985 2025-12-04T09:16:48.2361287Z * [new tag] viable/strict/1761863094 -> viable/strict/1761863094 2025-12-04T09:16:48.2362828Z * [new tag] viable/strict/1761864590 -> viable/strict/1761864590 2025-12-04T09:16:48.2364416Z * [new tag] viable/strict/1761866675 -> viable/strict/1761866675 2025-12-04T09:16:48.2366206Z * [new tag] viable/strict/1761868178 -> viable/strict/1761868178 2025-12-04T09:16:48.2367940Z * [new tag] viable/strict/1761871111 -> viable/strict/1761871111 2025-12-04T09:16:48.2369654Z * [new tag] viable/strict/1761873126 -> viable/strict/1761873126 2025-12-04T09:16:48.2371319Z * [new tag] viable/strict/1761875714 -> viable/strict/1761875714 2025-12-04T09:16:48.2373093Z * [new tag] viable/strict/1761878924 -> viable/strict/1761878924 2025-12-04T09:16:48.2375953Z * [new tag] viable/strict/1761881727 -> viable/strict/1761881727 2025-12-04T09:16:48.2376278Z * [new tag] viable/strict/1761882959 -> viable/strict/1761882959 2025-12-04T09:16:48.2378327Z * [new tag] viable/strict/1761886268 -> viable/strict/1761886268 2025-12-04T09:16:48.2379739Z * [new tag] viable/strict/1761893641 -> viable/strict/1761893641 2025-12-04T09:16:48.2381336Z * [new tag] viable/strict/1761931517 -> viable/strict/1761931517 2025-12-04T09:16:48.2382937Z * [new tag] viable/strict/1761933080 -> viable/strict/1761933080 2025-12-04T09:16:48.2384659Z * [new tag] viable/strict/1761935217 -> viable/strict/1761935217 2025-12-04T09:16:48.2386376Z * [new tag] viable/strict/1761938533 -> viable/strict/1761938533 2025-12-04T09:16:48.2387951Z * [new tag] viable/strict/1761940184 -> viable/strict/1761940184 2025-12-04T09:16:48.2389599Z * [new tag] viable/strict/1761942338 -> viable/strict/1761942338 2025-12-04T09:16:48.2391251Z * [new tag] viable/strict/1761946100 -> viable/strict/1761946100 2025-12-04T09:16:48.2392889Z * [new tag] viable/strict/1761947374 -> viable/strict/1761947374 2025-12-04T09:16:48.2394485Z * [new tag] viable/strict/1761950978 -> viable/strict/1761950978 2025-12-04T09:16:48.2396297Z * [new tag] viable/strict/1761957727 -> viable/strict/1761957727 2025-12-04T09:16:48.2397761Z * [new tag] viable/strict/1761959532 -> viable/strict/1761959532 2025-12-04T09:16:48.2399412Z * [new tag] viable/strict/1761965366 -> viable/strict/1761965366 2025-12-04T09:16:48.2401237Z * [new tag] viable/strict/1761968066 -> viable/strict/1761968066 2025-12-04T09:16:48.2402821Z * [new tag] viable/strict/1761969322 -> viable/strict/1761969322 2025-12-04T09:16:48.2404443Z * [new tag] viable/strict/1761974723 -> viable/strict/1761974723 2025-12-04T09:16:48.2406154Z * [new tag] viable/strict/1761981837 -> viable/strict/1761981837 2025-12-04T09:16:48.2407865Z * [new tag] viable/strict/1761985546 -> viable/strict/1761985546 2025-12-04T09:16:48.2409591Z * [new tag] viable/strict/1761987030 -> viable/strict/1761987030 2025-12-04T09:16:48.2411304Z * [new tag] viable/strict/1762003554 -> viable/strict/1762003554 2025-12-04T09:16:48.2412952Z * [new tag] viable/strict/1762021560 -> viable/strict/1762021560 2025-12-04T09:16:48.2414512Z * [new tag] viable/strict/1762032190 -> viable/strict/1762032190 2025-12-04T09:16:48.2416181Z * [new tag] viable/strict/1762040981 -> viable/strict/1762040981 2025-12-04T09:16:48.2417842Z * [new tag] viable/strict/1762048525 -> viable/strict/1762048525 2025-12-04T09:16:48.2419488Z * [new tag] viable/strict/1762104223 -> viable/strict/1762104223 2025-12-04T09:16:48.2421094Z * [new tag] viable/strict/1762105778 -> viable/strict/1762105778 2025-12-04T09:16:48.2422709Z * [new tag] viable/strict/1762115109 -> viable/strict/1762115109 2025-12-04T09:16:48.2424324Z * [new tag] viable/strict/1762125840 -> viable/strict/1762125840 2025-12-04T09:16:48.2425784Z * [new tag] viable/strict/1762127377 -> viable/strict/1762127377 2025-12-04T09:16:48.2427759Z * [new tag] viable/strict/1762134925 -> viable/strict/1762134925 2025-12-04T09:16:48.2429303Z * [new tag] viable/strict/1762138338 -> viable/strict/1762138338 2025-12-04T09:16:48.2430968Z * [new tag] viable/strict/1762148993 -> viable/strict/1762148993 2025-12-04T09:16:48.2432618Z * [new tag] viable/strict/1762152871 -> viable/strict/1762152871 2025-12-04T09:16:48.2434226Z * [new tag] viable/strict/1762156183 -> viable/strict/1762156183 2025-12-04T09:16:48.2436040Z * [new tag] viable/strict/1762163457 -> viable/strict/1762163457 2025-12-04T09:16:48.2437949Z * [new tag] viable/strict/1762165569 -> viable/strict/1762165569 2025-12-04T09:16:48.2439604Z * [new tag] viable/strict/1762169035 -> viable/strict/1762169035 2025-12-04T09:16:48.2441270Z * [new tag] viable/strict/1762174936 -> viable/strict/1762174936 2025-12-04T09:16:48.2442846Z * [new tag] viable/strict/1762194412 -> viable/strict/1762194412 2025-12-04T09:16:48.2444448Z * [new tag] viable/strict/1762195876 -> viable/strict/1762195876 2025-12-04T09:16:48.2446078Z * [new tag] viable/strict/1762197788 -> viable/strict/1762197788 2025-12-04T09:16:48.2447799Z * [new tag] viable/strict/1762199389 -> viable/strict/1762199389 2025-12-04T09:16:48.2449751Z * [new tag] viable/strict/1762206585 -> viable/strict/1762206585 2025-12-04T09:16:48.2451450Z * [new tag] viable/strict/1762210184 -> viable/strict/1762210184 2025-12-04T09:16:48.2452913Z * [new tag] viable/strict/1762218736 -> viable/strict/1762218736 2025-12-04T09:16:48.2454998Z * [new tag] viable/strict/1762224529 -> viable/strict/1762224529 2025-12-04T09:16:48.2456782Z * [new tag] viable/strict/1762227253 -> viable/strict/1762227253 2025-12-04T09:16:48.2458174Z * [new tag] viable/strict/1762228515 -> viable/strict/1762228515 2025-12-04T09:16:48.2459985Z * [new tag] viable/strict/1762230349 -> viable/strict/1762230349 2025-12-04T09:16:48.2461592Z * [new tag] viable/strict/1762231859 -> viable/strict/1762231859 2025-12-04T09:16:48.2463201Z * [new tag] viable/strict/1762233925 -> viable/strict/1762233925 2025-12-04T09:16:48.2464960Z * [new tag] viable/strict/1762237630 -> viable/strict/1762237630 2025-12-04T09:16:48.2466388Z * [new tag] viable/strict/1762253522 -> viable/strict/1762253522 2025-12-04T09:16:48.2468186Z * [new tag] viable/strict/1762278588 -> viable/strict/1762278588 2025-12-04T09:16:48.2469771Z * [new tag] viable/strict/1762284203 -> viable/strict/1762284203 2025-12-04T09:16:48.2471457Z * [new tag] viable/strict/1762289446 -> viable/strict/1762289446 2025-12-04T09:16:48.2473042Z * [new tag] viable/strict/1762291515 -> viable/strict/1762291515 2025-12-04T09:16:48.2474626Z * [new tag] viable/strict/1762295100 -> viable/strict/1762295100 2025-12-04T09:16:48.2476058Z * [new tag] viable/strict/1762296590 -> viable/strict/1762296590 2025-12-04T09:16:48.2477532Z * [new tag] viable/strict/1762300179 -> viable/strict/1762300179 2025-12-04T09:16:48.2478989Z * [new tag] viable/strict/1762303207 -> viable/strict/1762303207 2025-12-04T09:16:48.2480631Z * [new tag] viable/strict/1762386584 -> viable/strict/1762386584 2025-12-04T09:16:48.2482202Z * [new tag] viable/strict/1762391537 -> viable/strict/1762391537 2025-12-04T09:16:48.2483700Z * [new tag] viable/strict/1762394119 -> viable/strict/1762394119 2025-12-04T09:16:48.2485581Z * [new tag] viable/strict/1762397437 -> viable/strict/1762397437 2025-12-04T09:16:48.2487150Z * [new tag] viable/strict/1762400256 -> viable/strict/1762400256 2025-12-04T09:16:48.2488927Z * [new tag] viable/strict/1762401469 -> viable/strict/1762401469 2025-12-04T09:16:48.2490836Z * [new tag] viable/strict/1762408195 -> viable/strict/1762408195 2025-12-04T09:16:48.2492414Z * [new tag] viable/strict/1762410411 -> viable/strict/1762410411 2025-12-04T09:16:48.2494026Z * [new tag] viable/strict/1762417613 -> viable/strict/1762417613 2025-12-04T09:16:48.2495581Z * [new tag] viable/strict/1762419198 -> viable/strict/1762419198 2025-12-04T09:16:48.2497255Z * [new tag] viable/strict/1762422656 -> viable/strict/1762422656 2025-12-04T09:16:48.2499202Z * [new tag] viable/strict/1762424746 -> viable/strict/1762424746 2025-12-04T09:16:48.2500828Z * [new tag] viable/strict/1762446386 -> viable/strict/1762446386 2025-12-04T09:16:48.2502503Z * [new tag] viable/strict/1762449912 -> viable/strict/1762449912 2025-12-04T09:16:48.2504161Z * [new tag] viable/strict/1762457031 -> viable/strict/1762457031 2025-12-04T09:16:48.2505748Z * [new tag] viable/strict/1762462441 -> viable/strict/1762462441 2025-12-04T09:16:48.2507415Z * [new tag] viable/strict/1762467909 -> viable/strict/1762467909 2025-12-04T09:16:48.2509074Z * [new tag] viable/strict/1762471493 -> viable/strict/1762471493 2025-12-04T09:16:48.2510755Z * [new tag] viable/strict/1762475990 -> viable/strict/1762475990 2025-12-04T09:16:48.2512402Z * [new tag] viable/strict/1762477933 -> viable/strict/1762477933 2025-12-04T09:16:48.2513986Z * [new tag] viable/strict/1762491053 -> viable/strict/1762491053 2025-12-04T09:16:48.2515721Z * [new tag] viable/strict/1762493118 -> viable/strict/1762493118 2025-12-04T09:16:48.2517247Z * [new tag] viable/strict/1762498442 -> viable/strict/1762498442 2025-12-04T09:16:48.2518801Z * [new tag] viable/strict/1762501778 -> viable/strict/1762501778 2025-12-04T09:16:48.2520437Z * [new tag] viable/strict/1762504001 -> viable/strict/1762504001 2025-12-04T09:16:48.2522124Z * [new tag] viable/strict/1762505583 -> viable/strict/1762505583 2025-12-04T09:16:48.2523802Z * [new tag] viable/strict/1762507523 -> viable/strict/1762507523 2025-12-04T09:16:48.2525433Z * [new tag] viable/strict/1762511140 -> viable/strict/1762511140 2025-12-04T09:16:48.2527223Z * [new tag] viable/strict/1762512632 -> viable/strict/1762512632 2025-12-04T09:16:48.2529046Z * [new tag] viable/strict/1762520467 -> viable/strict/1762520467 2025-12-04T09:16:48.2530628Z * [new tag] viable/strict/1762522016 -> viable/strict/1762522016 2025-12-04T09:16:48.2532213Z * [new tag] viable/strict/1762530591 -> viable/strict/1762530591 2025-12-04T09:16:48.2533823Z * [new tag] viable/strict/1762543405 -> viable/strict/1762543405 2025-12-04T09:16:48.2535215Z * [new tag] viable/strict/1762544998 -> viable/strict/1762544998 2025-12-04T09:16:48.2536984Z * [new tag] viable/strict/1762552182 -> viable/strict/1762552182 2025-12-04T09:16:48.2538572Z * [new tag] viable/strict/1762554297 -> viable/strict/1762554297 2025-12-04T09:16:48.2540116Z * [new tag] viable/strict/1762559381 -> viable/strict/1762559381 2025-12-04T09:16:48.2541808Z * [new tag] viable/strict/1762562222 -> viable/strict/1762562222 2025-12-04T09:16:48.2543479Z * [new tag] viable/strict/1762564319 -> viable/strict/1762564319 2025-12-04T09:16:48.2544968Z * [new tag] viable/strict/1762566904 -> viable/strict/1762566904 2025-12-04T09:16:48.2546535Z * [new tag] viable/strict/1762569781 -> viable/strict/1762569781 2025-12-04T09:16:48.2548106Z * [new tag] viable/strict/1762575940 -> viable/strict/1762575940 2025-12-04T09:16:48.2549764Z * [new tag] viable/strict/1762580974 -> viable/strict/1762580974 2025-12-04T09:16:48.2551385Z * [new tag] viable/strict/1762583185 -> viable/strict/1762583185 2025-12-04T09:16:48.2553018Z * [new tag] viable/strict/1762586647 -> viable/strict/1762586647 2025-12-04T09:16:48.2554677Z * [new tag] viable/strict/1762588183 -> viable/strict/1762588183 2025-12-04T09:16:48.2556283Z * [new tag] viable/strict/1762593886 -> viable/strict/1762593886 2025-12-04T09:16:48.2557887Z * [new tag] viable/strict/1762650743 -> viable/strict/1762650743 2025-12-04T09:16:48.2560032Z * [new tag] viable/strict/1762653328 -> viable/strict/1762653328 2025-12-04T09:16:48.2561605Z * [new tag] viable/strict/1762659342 -> viable/strict/1762659342 2025-12-04T09:16:48.2563215Z * [new tag] viable/strict/1762662360 -> viable/strict/1762662360 2025-12-04T09:16:48.2564813Z * [new tag] viable/strict/1762667377 -> viable/strict/1762667377 2025-12-04T09:16:48.2566457Z * [new tag] viable/strict/1762671090 -> viable/strict/1762671090 2025-12-04T09:16:48.2568156Z * [new tag] viable/strict/1762680284 -> viable/strict/1762680284 2025-12-04T09:16:48.2569869Z * [new tag] viable/strict/1762683900 -> viable/strict/1762683900 2025-12-04T09:16:48.2571537Z * [new tag] viable/strict/1762705541 -> viable/strict/1762705541 2025-12-04T09:16:48.2573142Z * [new tag] viable/strict/1762709004 -> viable/strict/1762709004 2025-12-04T09:16:48.2575065Z * [new tag] viable/strict/1762746004 -> viable/strict/1762746004 2025-12-04T09:16:48.2576601Z * [new tag] viable/strict/1762748799 -> viable/strict/1762748799 2025-12-04T09:16:48.2578220Z * [new tag] viable/strict/1762759504 -> viable/strict/1762759504 2025-12-04T09:16:48.2579970Z * [new tag] viable/strict/1762760973 -> viable/strict/1762760973 2025-12-04T09:16:48.2581647Z * [new tag] viable/strict/1762775374 -> viable/strict/1762775374 2025-12-04T09:16:48.2583284Z * [new tag] viable/strict/1762777661 -> viable/strict/1762777661 2025-12-04T09:16:48.2584889Z * [new tag] viable/strict/1762779774 -> viable/strict/1762779774 2025-12-04T09:16:48.2586647Z * [new tag] viable/strict/1762781259 -> viable/strict/1762781259 2025-12-04T09:16:48.2588373Z * [new tag] viable/strict/1762793628 -> viable/strict/1762793628 2025-12-04T09:16:48.2590085Z * [new tag] viable/strict/1762800711 -> viable/strict/1762800711 2025-12-04T09:16:48.2591705Z * [new tag] viable/strict/1762809894 -> viable/strict/1762809894 2025-12-04T09:16:48.2593298Z * [new tag] viable/strict/1762811384 -> viable/strict/1762811384 2025-12-04T09:16:48.2595072Z * [new tag] viable/strict/1762813841 -> viable/strict/1762813841 2025-12-04T09:16:48.2596764Z * [new tag] viable/strict/1762815047 -> viable/strict/1762815047 2025-12-04T09:16:48.2598534Z * [new tag] viable/strict/1762817094 -> viable/strict/1762817094 2025-12-04T09:16:48.2600171Z * [new tag] viable/strict/1762818582 -> viable/strict/1762818582 2025-12-04T09:16:48.2601925Z * [new tag] viable/strict/1762821623 -> viable/strict/1762821623 2025-12-04T09:16:48.2603328Z * [new tag] viable/strict/1762823531 -> viable/strict/1762823531 2025-12-04T09:16:48.2605038Z * [new tag] viable/strict/1762849583 -> viable/strict/1762849583 2025-12-04T09:16:48.2606682Z * [new tag] viable/strict/1762851200 -> viable/strict/1762851200 2025-12-04T09:16:48.2608450Z * [new tag] viable/strict/1762854603 -> viable/strict/1762854603 2025-12-04T09:16:48.2610120Z * [new tag] viable/strict/1762858276 -> viable/strict/1762858276 2025-12-04T09:16:48.2611827Z * [new tag] viable/strict/1762860891 -> viable/strict/1762860891 2025-12-04T09:16:48.2614028Z * [new tag] viable/strict/1762866174 -> viable/strict/1762866174 2025-12-04T09:16:48.2615686Z * [new tag] viable/strict/1762867653 -> viable/strict/1762867653 2025-12-04T09:16:48.2617283Z * [new tag] viable/strict/1762872669 -> viable/strict/1762872669 2025-12-04T09:16:48.2618723Z * [new tag] viable/strict/1762878380 -> viable/strict/1762878380 2025-12-04T09:16:48.2620362Z * [new tag] viable/strict/1762889003 -> viable/strict/1762889003 2025-12-04T09:16:48.2622153Z * [new tag] viable/strict/1762890589 -> viable/strict/1762890589 2025-12-04T09:16:48.2623791Z * [new tag] viable/strict/1762892743 -> viable/strict/1762892743 2025-12-04T09:16:48.2625445Z * [new tag] viable/strict/1762894271 -> viable/strict/1762894271 2025-12-04T09:16:48.2626926Z * [new tag] viable/strict/1762896287 -> viable/strict/1762896287 2025-12-04T09:16:48.2628656Z * [new tag] viable/strict/1762915871 -> viable/strict/1762915871 2025-12-04T09:16:48.2630248Z * [new tag] viable/strict/1762918569 -> viable/strict/1762918569 2025-12-04T09:16:48.2631742Z * [new tag] viable/strict/1762919776 -> viable/strict/1762919776 2025-12-04T09:16:48.2633429Z * [new tag] viable/strict/1762923072 -> viable/strict/1762923072 2025-12-04T09:16:48.2635187Z * [new tag] viable/strict/1762928826 -> viable/strict/1762928826 2025-12-04T09:16:48.2637113Z * [new tag] viable/strict/1762930451 -> viable/strict/1762930451 2025-12-04T09:16:48.2638790Z * [new tag] viable/strict/1762933780 -> viable/strict/1762933780 2025-12-04T09:16:48.2640527Z * [new tag] viable/strict/1762937638 -> viable/strict/1762937638 2025-12-04T09:16:48.2642265Z * [new tag] viable/strict/1762939545 -> viable/strict/1762939545 2025-12-04T09:16:48.2643889Z * [new tag] viable/strict/1762962692 -> viable/strict/1762962692 2025-12-04T09:16:48.2645544Z * [new tag] viable/strict/1762979143 -> viable/strict/1762979143 2025-12-04T09:16:48.2647149Z * [new tag] viable/strict/1762984188 -> viable/strict/1762984188 2025-12-04T09:16:48.2648788Z * [new tag] viable/strict/1762986306 -> viable/strict/1762986306 2025-12-04T09:16:48.2650547Z * [new tag] viable/strict/1762989903 -> viable/strict/1762989903 2025-12-04T09:16:48.2652488Z * [new tag] viable/strict/1762991377 -> viable/strict/1762991377 2025-12-04T09:16:48.2654119Z * [new tag] viable/strict/1762998921 -> viable/strict/1762998921 2025-12-04T09:16:48.2655855Z * [new tag] viable/strict/1763002287 -> viable/strict/1763002287 2025-12-04T09:16:48.2657452Z * [new tag] viable/strict/1763016840 -> viable/strict/1763016840 2025-12-04T09:16:48.2659053Z * [new tag] viable/strict/1763020180 -> viable/strict/1763020180 2025-12-04T09:16:48.2660686Z * [new tag] viable/strict/1763027421 -> viable/strict/1763027421 2025-12-04T09:16:48.2662330Z * [new tag] viable/strict/1763031120 -> viable/strict/1763031120 2025-12-04T09:16:48.2663942Z * [new tag] viable/strict/1763036861 -> viable/strict/1763036861 2025-12-04T09:16:48.2665631Z * [new tag] viable/strict/1763038993 -> viable/strict/1763038993 2025-12-04T09:16:48.2667760Z * [new tag] viable/strict/1763054703 -> viable/strict/1763054703 2025-12-04T09:16:48.2669240Z * [new tag] viable/strict/1763067061 -> viable/strict/1763067061 2025-12-04T09:16:48.2670834Z * [new tag] viable/strict/1763070847 -> viable/strict/1763070847 2025-12-04T09:16:48.2672532Z * [new tag] viable/strict/1763072706 -> viable/strict/1763072706 2025-12-04T09:16:48.2674174Z * [new tag] viable/strict/1763076302 -> viable/strict/1763076302 2025-12-04T09:16:48.2675797Z * [new tag] viable/strict/1763080816 -> viable/strict/1763080816 2025-12-04T09:16:48.2677451Z * [new tag] viable/strict/1763082732 -> viable/strict/1763082732 2025-12-04T09:16:48.2679035Z * [new tag] viable/strict/1763085329 -> viable/strict/1763085329 2025-12-04T09:16:48.2680654Z * [new tag] viable/strict/1763088623 -> viable/strict/1763088623 2025-12-04T09:16:48.2682373Z * [new tag] viable/strict/1763091402 -> viable/strict/1763091402 2025-12-04T09:16:48.2683984Z * [new tag] viable/strict/1763092602 -> viable/strict/1763092602 2025-12-04T09:16:48.2685612Z * [new tag] viable/strict/1763094355 -> viable/strict/1763094355 2025-12-04T09:16:48.2687229Z * [new tag] viable/strict/1763099390 -> viable/strict/1763099390 2025-12-04T09:16:48.2689065Z * [new tag] viable/strict/1763101608 -> viable/strict/1763101608 2025-12-04T09:16:48.2690674Z * [new tag] viable/strict/1763105102 -> viable/strict/1763105102 2025-12-04T09:16:48.2692328Z * [new tag] viable/strict/1763112347 -> viable/strict/1763112347 2025-12-04T09:16:48.2693912Z * [new tag] viable/strict/1763119471 -> viable/strict/1763119471 2025-12-04T09:16:48.2695484Z * [new tag] viable/strict/1763126835 -> viable/strict/1763126835 2025-12-04T09:16:48.2696794Z * [new tag] viable/strict/1763149779 -> viable/strict/1763149779 2025-12-04T09:16:48.2698437Z * [new tag] viable/strict/1763164178 -> viable/strict/1763164178 2025-12-04T09:16:48.2700108Z * [new tag] viable/strict/1763167104 -> viable/strict/1763167104 2025-12-04T09:16:48.2701695Z * [new tag] viable/strict/1763169132 -> viable/strict/1763169132 2025-12-04T09:16:48.2703418Z * [new tag] viable/strict/1763171708 -> viable/strict/1763171708 2025-12-04T09:16:48.2705128Z * [new tag] viable/strict/1763174759 -> viable/strict/1763174759 2025-12-04T09:16:48.2706815Z * [new tag] viable/strict/1763180744 -> viable/strict/1763180744 2025-12-04T09:16:48.2708379Z * [new tag] viable/strict/1763182227 -> viable/strict/1763182227 2025-12-04T09:16:48.2710022Z * [new tag] viable/strict/1763184309 -> viable/strict/1763184309 2025-12-04T09:16:48.2712064Z * [new tag] viable/strict/1763187991 -> viable/strict/1763187991 2025-12-04T09:16:48.2713701Z * [new tag] viable/strict/1763191445 -> viable/strict/1763191445 2025-12-04T09:16:48.2715491Z * [new tag] viable/strict/1763195152 -> viable/strict/1763195152 2025-12-04T09:16:48.2716978Z * [new tag] viable/strict/1763205769 -> viable/strict/1763205769 2025-12-04T09:16:48.2718606Z * [new tag] viable/strict/1763246990 -> viable/strict/1763246990 2025-12-04T09:16:48.2720371Z * [new tag] viable/strict/1763261578 -> viable/strict/1763261578 2025-12-04T09:16:48.2721778Z * [new tag] viable/strict/1763286573 -> viable/strict/1763286573 2025-12-04T09:16:48.2723249Z * [new tag] viable/strict/1763292167 -> viable/strict/1763292167 2025-12-04T09:16:48.2724956Z * [new tag] viable/strict/1763333386 -> viable/strict/1763333386 2025-12-04T09:16:48.2726608Z * [new tag] viable/strict/1763340082 -> viable/strict/1763340082 2025-12-04T09:16:48.2729022Z * [new tag] viable/strict/1763364324 -> viable/strict/1763364324 2025-12-04T09:16:48.2730594Z * [new tag] viable/strict/1763371569 -> viable/strict/1763371569 2025-12-04T09:16:48.2732186Z * [new tag] viable/strict/1763373067 -> viable/strict/1763373067 2025-12-04T09:16:48.2733790Z * [new tag] viable/strict/1763375157 -> viable/strict/1763375157 2025-12-04T09:16:48.2735573Z * [new tag] viable/strict/1763382462 -> viable/strict/1763382462 2025-12-04T09:16:48.2738673Z * [new tag] viable/strict/1763394661 -> viable/strict/1763394661 2025-12-04T09:16:48.2740422Z * [new tag] viable/strict/1763396797 -> viable/strict/1763396797 2025-12-04T09:16:48.2742187Z * [new tag] viable/strict/1763398542 -> viable/strict/1763398542 2025-12-04T09:16:48.2743750Z * [new tag] viable/strict/1763401807 -> viable/strict/1763401807 2025-12-04T09:16:48.2745201Z * [new tag] viable/strict/1763414698 -> viable/strict/1763414698 2025-12-04T09:16:48.2746842Z * [new tag] viable/strict/1763419807 -> viable/strict/1763419807 2025-12-04T09:16:48.2748454Z * [new tag] viable/strict/1763426369 -> viable/strict/1763426369 2025-12-04T09:16:48.2750065Z * [new tag] viable/strict/1763428331 -> viable/strict/1763428331 2025-12-04T09:16:48.2751732Z * [new tag] viable/strict/1763430922 -> viable/strict/1763430922 2025-12-04T09:16:48.2753206Z * [new tag] viable/strict/1763434184 -> viable/strict/1763434184 2025-12-04T09:16:48.2754854Z * [new tag] viable/strict/1763439973 -> viable/strict/1763439973 2025-12-04T09:16:48.2756631Z * [new tag] viable/strict/1763444995 -> viable/strict/1763444995 2025-12-04T09:16:48.2758203Z * [new tag] viable/strict/1763447206 -> viable/strict/1763447206 2025-12-04T09:16:48.2759883Z * [new tag] viable/strict/1763448826 -> viable/strict/1763448826 2025-12-04T09:16:48.2761548Z * [new tag] viable/strict/1763450717 -> viable/strict/1763450717 2025-12-04T09:16:48.2763137Z * [new tag] viable/strict/1763452183 -> viable/strict/1763452183 2025-12-04T09:16:48.2764812Z * [new tag] viable/strict/1763457945 -> viable/strict/1763457945 2025-12-04T09:16:48.2766420Z * [new tag] viable/strict/1763459439 -> viable/strict/1763459439 2025-12-04T09:16:48.2767965Z * [new tag] viable/strict/1763461556 -> viable/strict/1763461556 2025-12-04T09:16:48.2769601Z * [new tag] viable/strict/1763463103 -> viable/strict/1763463103 2025-12-04T09:16:48.2771294Z * [new tag] viable/strict/1763465100 -> viable/strict/1763465100 2025-12-04T09:16:48.2772740Z * [new tag] viable/strict/1763468866 -> viable/strict/1763468866 2025-12-04T09:16:48.2774651Z * [new tag] viable/strict/1763493823 -> viable/strict/1763493823 2025-12-04T09:16:48.2776106Z * [new tag] viable/strict/1763496249 -> viable/strict/1763496249 2025-12-04T09:16:48.2777749Z * [new tag] viable/strict/1763502620 -> viable/strict/1763502620 2025-12-04T09:16:48.2779422Z * [new tag] viable/strict/1763504715 -> viable/strict/1763504715 2025-12-04T09:16:48.2781060Z * [new tag] viable/strict/1763506208 -> viable/strict/1763506208 2025-12-04T09:16:48.2782638Z * [new tag] viable/strict/1763520590 -> viable/strict/1763520590 2025-12-04T09:16:48.2784373Z * [new tag] viable/strict/1763523357 -> viable/strict/1763523357 2025-12-04T09:16:48.2786007Z * [new tag] viable/strict/1763529922 -> viable/strict/1763529922 2025-12-04T09:16:48.2787747Z * [new tag] viable/strict/1763531408 -> viable/strict/1763531408 2025-12-04T09:16:48.2789360Z * [new tag] viable/strict/1763533622 -> viable/strict/1763533622 2025-12-04T09:16:48.2790972Z * [new tag] viable/strict/1763538576 -> viable/strict/1763538576 2025-12-04T09:16:48.2792688Z * [new tag] viable/strict/1763545823 -> viable/strict/1763545823 2025-12-04T09:16:48.2794090Z * [new tag] viable/strict/1763547951 -> viable/strict/1763547951 2025-12-04T09:16:48.2795737Z * [new tag] viable/strict/1763551477 -> viable/strict/1763551477 2025-12-04T09:16:48.2797391Z * [new tag] viable/strict/1763552982 -> viable/strict/1763552982 2025-12-04T09:16:48.2799053Z * [new tag] viable/strict/1763594698 -> viable/strict/1763594698 2025-12-04T09:16:48.2800620Z * [new tag] viable/strict/1763596178 -> viable/strict/1763596178 2025-12-04T09:16:48.2802271Z * [new tag] viable/strict/1763599155 -> viable/strict/1763599155 2025-12-04T09:16:48.2803862Z * [new tag] viable/strict/1763603717 -> viable/strict/1763603717 2025-12-04T09:16:48.2805553Z * [new tag] viable/strict/1763606923 -> viable/strict/1763606923 2025-12-04T09:16:48.2807140Z * [new tag] viable/strict/1763609715 -> viable/strict/1763609715 2025-12-04T09:16:48.2808883Z * [new tag] viable/strict/1763612757 -> viable/strict/1763612757 2025-12-04T09:16:48.2810664Z * [new tag] viable/strict/1763616325 -> viable/strict/1763616325 2025-12-04T09:16:48.2812187Z * [new tag] viable/strict/1763623509 -> viable/strict/1763623509 2025-12-04T09:16:48.2813895Z * [new tag] viable/strict/1763624984 -> viable/strict/1763624984 2025-12-04T09:16:48.2815684Z * [new tag] viable/strict/1763628796 -> viable/strict/1763628796 2025-12-04T09:16:48.2817106Z * [new tag] viable/strict/1763634343 -> viable/strict/1763634343 2025-12-04T09:16:48.2818679Z * [new tag] viable/strict/1763635867 -> viable/strict/1763635867 2025-12-04T09:16:48.2820498Z * [new tag] viable/strict/1763639382 -> viable/strict/1763639382 2025-12-04T09:16:48.2822088Z * [new tag] viable/strict/1763646626 -> viable/strict/1763646626 2025-12-04T09:16:48.2823848Z * [new tag] viable/strict/1763655997 -> viable/strict/1763655997 2025-12-04T09:16:48.2825478Z * [new tag] viable/strict/1763659444 -> viable/strict/1763659444 2025-12-04T09:16:48.2827067Z * [new tag] viable/strict/1763660992 -> viable/strict/1763660992 2025-12-04T09:16:48.2828656Z * [new tag] viable/strict/1763663201 -> viable/strict/1763663201 2025-12-04T09:16:48.2830293Z * [new tag] viable/strict/1763670362 -> viable/strict/1763670362 2025-12-04T09:16:48.2831727Z * [new tag] viable/strict/1763675378 -> viable/strict/1763675378 2025-12-04T09:16:48.2833397Z * [new tag] viable/strict/1763693343 -> viable/strict/1763693343 2025-12-04T09:16:48.2834954Z * [new tag] viable/strict/1763696088 -> viable/strict/1763696088 2025-12-04T09:16:48.2837067Z * [new tag] viable/strict/1763697343 -> viable/strict/1763697343 2025-12-04T09:16:48.2838643Z * [new tag] viable/strict/1763699165 -> viable/strict/1763699165 2025-12-04T09:16:48.2840243Z * [new tag] viable/strict/1763700660 -> viable/strict/1763700660 2025-12-04T09:16:48.2841778Z * [new tag] viable/strict/1763704209 -> viable/strict/1763704209 2025-12-04T09:16:48.2843367Z * [new tag] viable/strict/1763706411 -> viable/strict/1763706411 2025-12-04T09:16:48.2844963Z * [new tag] viable/strict/1763708082 -> viable/strict/1763708082 2025-12-04T09:16:48.2846458Z * [new tag] viable/strict/1763711381 -> viable/strict/1763711381 2025-12-04T09:16:48.2848097Z * [new tag] viable/strict/1763713593 -> viable/strict/1763713593 2025-12-04T09:16:48.2849728Z * [new tag] viable/strict/1763715201 -> viable/strict/1763715201 2025-12-04T09:16:48.2851284Z * [new tag] viable/strict/1763733017 -> viable/strict/1763733017 2025-12-04T09:16:48.2852901Z * [new tag] viable/strict/1763735108 -> viable/strict/1763735108 2025-12-04T09:16:48.2854488Z * [new tag] viable/strict/1763749579 -> viable/strict/1763749579 2025-12-04T09:16:48.2856105Z * [new tag] viable/strict/1763751113 -> viable/strict/1763751113 2025-12-04T09:16:48.2857700Z * [new tag] viable/strict/1763753035 -> viable/strict/1763753035 2025-12-04T09:16:48.2859381Z * [new tag] viable/strict/1763754578 -> viable/strict/1763754578 2025-12-04T09:16:48.2860939Z * [new tag] viable/strict/1763756748 -> viable/strict/1763756748 2025-12-04T09:16:48.2862617Z * [new tag] viable/strict/1763758205 -> viable/strict/1763758205 2025-12-04T09:16:48.2864095Z * [new tag] viable/strict/1763764050 -> viable/strict/1763764050 2025-12-04T09:16:48.2865794Z * [new tag] viable/strict/1763771887 -> viable/strict/1763771887 2025-12-04T09:16:48.2867505Z * [new tag] viable/strict/1763773920 -> viable/strict/1763773920 2025-12-04T09:16:48.2869128Z * [new tag] viable/strict/1763776501 -> viable/strict/1763776501 2025-12-04T09:16:48.2870703Z * [new tag] viable/strict/1763779437 -> viable/strict/1763779437 2025-12-04T09:16:48.2872494Z * [new tag] viable/strict/1763781038 -> viable/strict/1763781038 2025-12-04T09:16:48.2874154Z * [new tag] viable/strict/1763782245 -> viable/strict/1763782245 2025-12-04T09:16:48.2875665Z * [new tag] viable/strict/1763785568 -> viable/strict/1763785568 2025-12-04T09:16:48.2877258Z * [new tag] viable/strict/1763787006 -> viable/strict/1763787006 2025-12-04T09:16:48.2879436Z * [new tag] viable/strict/1763789103 -> viable/strict/1763789103 2025-12-04T09:16:48.2880963Z * [new tag] viable/strict/1763790578 -> viable/strict/1763790578 2025-12-04T09:16:48.2882629Z * [new tag] viable/strict/1763796275 -> viable/strict/1763796275 2025-12-04T09:16:48.2884377Z * [new tag] viable/strict/1763801465 -> viable/strict/1763801465 2025-12-04T09:16:48.2885939Z * [new tag] viable/strict/1763803522 -> viable/strict/1763803522 2025-12-04T09:16:48.2887614Z * [new tag] viable/strict/1763808581 -> viable/strict/1763808581 2025-12-04T09:16:48.2889420Z * [new tag] viable/strict/1763840977 -> viable/strict/1763840977 2025-12-04T09:16:48.2890949Z * [new tag] viable/strict/1763846659 -> viable/strict/1763846659 2025-12-04T09:16:48.2892517Z * [new tag] viable/strict/1763872065 -> viable/strict/1763872065 2025-12-04T09:16:48.2894171Z * [new tag] viable/strict/1763873648 -> viable/strict/1763873648 2025-12-04T09:16:48.2895806Z * [new tag] viable/strict/1763875506 -> viable/strict/1763875506 2025-12-04T09:16:48.2897258Z * [new tag] viable/strict/1763889904 -> viable/strict/1763889904 2025-12-04T09:16:48.2898854Z * [new tag] viable/strict/1763930999 -> viable/strict/1763930999 2025-12-04T09:16:48.2900515Z * [new tag] viable/strict/1763944964 -> viable/strict/1763944964 2025-12-04T09:16:48.2902011Z * [new tag] viable/strict/1763958474 -> viable/strict/1763958474 2025-12-04T09:16:48.2903581Z * [new tag] viable/strict/1763967263 -> viable/strict/1763967263 2025-12-04T09:16:48.2905175Z * [new tag] viable/strict/1763972803 -> viable/strict/1763972803 2025-12-04T09:16:48.2906753Z * [new tag] viable/strict/1763976376 -> viable/strict/1763976376 2025-12-04T09:16:48.2908366Z * [new tag] viable/strict/1763989404 -> viable/strict/1763989404 2025-12-04T09:16:48.2909945Z * [new tag] viable/strict/1763990887 -> viable/strict/1763990887 2025-12-04T09:16:48.2911501Z * [new tag] viable/strict/1764019919 -> viable/strict/1764019919 2025-12-04T09:16:48.2913178Z * [new tag] viable/strict/1764023134 -> viable/strict/1764023134 2025-12-04T09:16:48.2914716Z * [new tag] viable/strict/1764024593 -> viable/strict/1764024593 2025-12-04T09:16:48.2916277Z * [new tag] viable/strict/1764026706 -> viable/strict/1764026706 2025-12-04T09:16:48.2918115Z * [new tag] viable/strict/1764031139 -> viable/strict/1764031139 2025-12-04T09:16:48.2919775Z * [new tag] viable/strict/1764033131 -> viable/strict/1764033131 2025-12-04T09:16:48.2921209Z * [new tag] viable/strict/1764035725 -> viable/strict/1764035725 2025-12-04T09:16:48.2922644Z * [new tag] viable/strict/1764624265 -> viable/strict/1764624265 2025-12-04T09:16:48.2924092Z * [new tag] viable/strict/1764631514 -> viable/strict/1764631514 2025-12-04T09:16:48.2925515Z * [new tag] viable/strict/1764632987 -> viable/strict/1764632987 2025-12-04T09:16:48.2926938Z * [new tag] viable/strict/1764636063 -> viable/strict/1764636063 2025-12-04T09:16:48.2928599Z * [new tag] viable/strict/1764643975 -> viable/strict/1764643975 2025-12-04T09:16:48.2930038Z * [new tag] viable/strict/1764646859 -> viable/strict/1764646859 2025-12-04T09:16:48.2931606Z * [new tag] viable/strict/1764653120 -> viable/strict/1764653120 2025-12-04T09:16:48.2932937Z * [new tag] viable/strict/1764654632 -> viable/strict/1764654632 2025-12-04T09:16:48.2934357Z * [new tag] viable/strict/1764656821 -> viable/strict/1764656821 2025-12-04T09:16:48.2936031Z * [new tag] viable/strict/1764658557 -> viable/strict/1764658557 2025-12-04T09:16:48.2937450Z * [new tag] viable/strict/1764660333 -> viable/strict/1764660333 2025-12-04T09:16:48.2938830Z * [new tag] viable/strict/1764661812 -> viable/strict/1764661812 2025-12-04T09:16:48.2940281Z * [new tag] viable/strict/1764664023 -> viable/strict/1764664023 2025-12-04T09:16:48.2941717Z * [new tag] viable/strict/1764669150 -> viable/strict/1764669150 2025-12-04T09:16:48.2943150Z * [new tag] viable/strict/1764680709 -> viable/strict/1764680709 2025-12-04T09:16:48.2944601Z * [new tag] viable/strict/1764687619 -> viable/strict/1764687619 2025-12-04T09:16:48.2946047Z * [new tag] viable/strict/1764696355 -> viable/strict/1764696355 2025-12-04T09:16:48.2947465Z * [new tag] viable/strict/1764701767 -> viable/strict/1764701767 2025-12-04T09:16:48.2948974Z * [new tag] viable/strict/1764710768 -> viable/strict/1764710768 2025-12-04T09:16:48.2950357Z * [new tag] viable/strict/1764716202 -> viable/strict/1764716202 2025-12-04T09:16:48.2951810Z * [new tag] viable/strict/1764793566 -> viable/strict/1764793566 2025-12-04T09:16:48.2953273Z * [new tag] viable/strict/1764797093 -> viable/strict/1764797093 2025-12-04T09:16:48.2954703Z * [new tag] viable/strict/1764800729 -> viable/strict/1764800729 2025-12-04T09:16:48.2956322Z * [new tag] whc_flight_1 -> whc_flight_1 2025-12-04T09:16:48.2957872Z * [new tag] whc_flight_2 -> whc_flight_2 2025-12-04T09:16:48.2959618Z * [new tag] whc_flight_4 -> whc_flight_4 2025-12-04T09:16:48.4144862Z [command]/usr/bin/git rev-parse --verify --quiet ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32^{object} 2025-12-04T09:16:48.4177254Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:16:48.4182478Z ##[endgroup] 2025-12-04T09:16:48.4182911Z ##[group]Determining the checkout info 2025-12-04T09:16:48.4184000Z ##[endgroup] 2025-12-04T09:16:48.4188453Z [command]/usr/bin/git sparse-checkout disable 2025-12-04T09:16:48.4231537Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-12-04T09:16:48.4266837Z ##[group]Checking out the ref 2025-12-04T09:16:48.4270060Z [command]/usr/bin/git checkout --progress --force ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:16:49.4536757Z Updating files: 65% (13240/20121) 2025-12-04T09:16:49.4620938Z Updating files: 66% (13280/20121) 2025-12-04T09:16:49.4703209Z Updating files: 67% (13482/20121) 2025-12-04T09:16:49.4787042Z Updating files: 68% (13683/20121) 2025-12-04T09:16:49.5003725Z Updating files: 69% (13884/20121) 2025-12-04T09:16:49.5337797Z Updating files: 70% (14085/20121) 2025-12-04T09:16:49.5407383Z Updating files: 71% (14286/20121) 2025-12-04T09:16:49.5501041Z Updating files: 72% (14488/20121) 2025-12-04T09:16:49.5720875Z Updating files: 73% (14689/20121) 2025-12-04T09:16:49.5995847Z Updating files: 74% (14890/20121) 2025-12-04T09:16:49.6563483Z Updating files: 75% (15091/20121) 2025-12-04T09:16:49.6742124Z Updating files: 76% (15292/20121) 2025-12-04T09:16:49.6908550Z Updating files: 77% (15494/20121) 2025-12-04T09:16:49.7148156Z Updating files: 78% (15695/20121) 2025-12-04T09:16:49.7449335Z Updating files: 79% (15896/20121) 2025-12-04T09:16:49.7807348Z Updating files: 80% (16097/20121) 2025-12-04T09:16:49.8130784Z Updating files: 81% (16299/20121) 2025-12-04T09:16:49.8389682Z Updating files: 82% (16500/20121) 2025-12-04T09:16:49.8579060Z Updating files: 83% (16701/20121) 2025-12-04T09:16:49.8755800Z Updating files: 84% (16902/20121) 2025-12-04T09:16:49.8953859Z Updating files: 85% (17103/20121) 2025-12-04T09:16:49.9148204Z Updating files: 86% (17305/20121) 2025-12-04T09:16:49.9325060Z Updating files: 87% (17506/20121) 2025-12-04T09:16:49.9472971Z Updating files: 88% (17707/20121) 2025-12-04T09:16:49.9646490Z Updating files: 89% (17908/20121) 2025-12-04T09:16:49.9857949Z Updating files: 90% (18109/20121) 2025-12-04T09:16:50.0010847Z Updating files: 91% (18311/20121) 2025-12-04T09:16:50.0205171Z Updating files: 92% (18512/20121) 2025-12-04T09:16:50.0427677Z Updating files: 93% (18713/20121) 2025-12-04T09:16:50.0669134Z Updating files: 94% (18914/20121) 2025-12-04T09:16:50.0883347Z Updating files: 95% (19115/20121) 2025-12-04T09:16:50.1076154Z Updating files: 96% (19317/20121) 2025-12-04T09:16:50.1279493Z Updating files: 97% (19518/20121) 2025-12-04T09:16:50.1605143Z Updating files: 98% (19719/20121) 2025-12-04T09:16:50.1818504Z Updating files: 99% (19920/20121) 2025-12-04T09:16:50.1818927Z Updating files: 100% (20121/20121) 2025-12-04T09:16:50.1819342Z Updating files: 100% (20121/20121), done. 2025-12-04T09:16:50.2107387Z Note: switching to 'ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32'. 2025-12-04T09:16:50.2107824Z 2025-12-04T09:16:50.2108059Z You are in 'detached HEAD' state. You can look around, make experimental 2025-12-04T09:16:50.2108585Z changes and commit them, and you can discard any commits you make in this 2025-12-04T09:16:50.2109125Z state without impacting any branches by switching back to a branch. 2025-12-04T09:16:50.2109435Z 2025-12-04T09:16:50.2109656Z If you want to create a new branch to retain commits you create, you may 2025-12-04T09:16:50.2110152Z do so (now or later) by using -c with the switch command. Example: 2025-12-04T09:16:50.2110432Z 2025-12-04T09:16:50.2110550Z git switch -c 2025-12-04T09:16:50.2110756Z 2025-12-04T09:16:50.2110875Z Or undo this operation with: 2025-12-04T09:16:50.2111058Z 2025-12-04T09:16:50.2111164Z git switch - 2025-12-04T09:16:50.2111342Z 2025-12-04T09:16:50.2111633Z Turn off this advice by setting config variable advice.detachedHead to false 2025-12-04T09:16:50.2111980Z 2025-12-04T09:16:50.2113471Z HEAD is now at ffd9b0fb435 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T09:16:50.2297527Z ##[endgroup] 2025-12-04T09:16:50.2298139Z ##[group]Setting up auth for fetching submodules 2025-12-04T09:16:50.2311865Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T09:16:50.2368872Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-12-04T09:16:50.2404741Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-12-04T09:16:50.2442218Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-12-04T09:16:50.2474568Z ##[endgroup] 2025-12-04T09:16:50.2475373Z ##[group]Fetching submodules 2025-12-04T09:16:50.2479204Z [command]/usr/bin/git submodule sync --recursive 2025-12-04T09:16:50.2897520Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-12-04T09:16:50.3300095Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-12-04T09:16:50.3302784Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-12-04T09:16:50.3306820Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-12-04T09:16:50.3310931Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-12-04T09:16:50.3314999Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-12-04T09:16:50.3319713Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-12-04T09:16:50.3323628Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-12-04T09:16:50.3328238Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-12-04T09:16:50.3332875Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-12-04T09:16:50.3338650Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-12-04T09:16:50.3342969Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-12-04T09:16:50.3347788Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-12-04T09:16:50.3352672Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-12-04T09:16:50.3357564Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-12-04T09:16:50.3362570Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-12-04T09:16:50.3369040Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-12-04T09:16:50.3377774Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-12-04T09:16:50.3383360Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-12-04T09:16:50.3388739Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:16:50.3393957Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-12-04T09:16:50.3399561Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-12-04T09:16:50.3405083Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-12-04T09:16:50.3411941Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-12-04T09:16:50.3416366Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-12-04T09:16:50.3421932Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-12-04T09:16:50.3427550Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-12-04T09:16:50.3433458Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-12-04T09:16:50.3440227Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-12-04T09:16:50.3446283Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-12-04T09:16:50.3452401Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-12-04T09:16:50.3458526Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-12-04T09:16:50.3465215Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-12-04T09:16:50.3471816Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-12-04T09:16:50.3481448Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-12-04T09:16:50.3487991Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-12-04T09:16:50.3494694Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-12-04T09:16:50.3501361Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-12-04T09:16:50.3542443Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-12-04T09:16:50.6360851Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-12-04T09:16:50.6361825Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-12-04T09:16:50.6362574Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-12-04T09:16:50.6399051Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-12-04T09:16:53.7030829Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-12-04T09:16:53.7031823Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-12-04T09:16:53.7032798Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-12-04T09:16:53.7033815Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-12-04T09:16:53.7035218Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-12-04T09:16:53.7036791Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-12-04T09:16:53.7037860Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-12-04T09:16:53.7038836Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-12-04T09:16:53.7039920Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-12-04T09:16:53.7040952Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-12-04T09:16:53.7041989Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-12-04T09:16:53.7043090Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-12-04T09:16:53.7044073Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-12-04T09:16:53.7045050Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-12-04T09:16:53.7046021Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-12-04T09:16:53.7046972Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-12-04T09:16:53.7048390Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-12-04T09:16:53.7774779Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-12-04T09:16:53.7949032Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-12-04T09:16:54.5174457Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-12-04T09:16:54.5176136Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-12-04T09:16:54.6184753Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-12-04T09:16:56.6566868Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-12-04T09:16:56.6568433Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-12-04T09:16:56.6569612Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-12-04T09:16:56.6570765Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-12-04T09:16:56.7567696Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-12-04T09:17:14.3541240Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-12-04T09:17:14.3542007Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-12-04T09:17:14.3542736Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-12-04T09:17:14.3543438Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-12-04T09:17:14.3544130Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-12-04T09:17:14.3758691Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-12-04T09:17:14.3932511Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-12-04T09:17:14.4070938Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-12-04T09:17:14.4422219Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-12-04T09:17:14.5497740Z Submodule path 'third_party/NVTX': checked out '3ebbc93ded7285963bff932c678fa367eb393ba6' 2025-12-04T09:17:14.6100689Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-12-04T09:17:15.6028338Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-12-04T09:17:15.8303710Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-12-04T09:17:15.8333075Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:17:15.8368175Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-12-04T09:17:21.3574291Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-12-04T09:17:21.3907121Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-12-04T09:17:21.8703850Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T09:17:21.9348286Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-12-04T09:17:22.0638404Z Submodule path 'third_party/cpuinfo': checked out 'f858c30bcb16f8effd5ff46996f0514539e17abc' 2025-12-04T09:17:22.1251765Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-12-04T09:17:22.9667527Z Submodule path 'third_party/cutlass': checked out 'f88806b1e31dfa579842638740216dd41fc6c588' 2025-12-04T09:17:23.1693512Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-12-04T09:17:23.1725082Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-12-04T09:17:23.1727974Z Submodule 'external/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:17:23.1733118Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:17:23.1737385Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-12-04T09:17:23.1741543Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-12-04T09:17:23.1745532Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:17:23.1749661Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-12-04T09:17:23.1786285Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-12-04T09:17:24.4423155Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-12-04T09:17:24.4424059Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-12-04T09:17:24.4424953Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-12-04T09:17:24.5424049Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-12-04T09:17:28.1242255Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-12-04T09:17:28.2242655Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-12-04T09:17:31.1697147Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-12-04T09:17:31.6437955Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T09:17:31.7651441Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-12-04T09:17:32.5826045Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-12-04T09:17:32.6401416Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:17:32.6566000Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-12-04T09:17:32.7973214Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-12-04T09:17:32.8928085Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-12-04T09:17:32.8952090Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:17:32.8955765Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:17:32.8991463Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-12-04T09:17:37.8080217Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-12-04T09:17:38.1425365Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-12-04T09:17:38.8643480Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-12-04T09:17:39.0521285Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-12-04T09:17:39.0906912Z Submodule path 'third_party/fmt': checked out '407c905e45ad75fc29bf0f9bb7c5c2fd3475976f' 2025-12-04T09:17:39.1401682Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-12-04T09:17:39.1761722Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-12-04T09:17:39.2326557Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:17:39.2513972Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-12-04T09:17:39.2541137Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-12-04T09:17:39.2571189Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-12-04T09:17:56.8220541Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-12-04T09:17:56.8504820Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-12-04T09:17:56.9531038Z Submodule path 'third_party/kineto': checked out '31f85df8fbd89c188f14ef10f1ec65379786b943' 2025-12-04T09:17:56.9554089Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:17:56.9557792Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:17:56.9561992Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:17:56.9596541Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-12-04T09:17:57.6639056Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-12-04T09:17:58.3126952Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-12-04T09:17:58.4224537Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-12-04T09:17:58.4249846Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:17:58.4252889Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:17:58.4256672Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:17:58.4260835Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:17:58.4265007Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:17:58.4269385Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:17:58.4273810Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:17:58.4278292Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:17:58.4283448Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:17:58.4318754Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-12-04T09:18:00.3866940Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-12-04T09:18:00.3869146Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp'... 2025-12-04T09:18:00.3870537Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-12-04T09:18:00.3871606Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-12-04T09:18:00.3872804Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-12-04T09:18:00.3873881Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-12-04T09:18:00.3874948Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-12-04T09:18:00.4867721Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-12-04T09:18:06.4420754Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-12-04T09:18:06.4680172Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-12-04T09:18:06.5146006Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-12-04T09:18:06.5334090Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-12-04T09:18:06.5357835Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:18:06.5398520Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-12-04T09:18:06.9815071Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-12-04T09:18:07.0073577Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-12-04T09:18:07.0636564Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:18:07.1931523Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-12-04T09:18:07.2161448Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-12-04T09:18:07.2415530Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-12-04T09:18:07.2438436Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:07.2442316Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:07.2478008Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T09:18:09.7585549Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T09:18:10.0542071Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-12-04T09:18:10.1130409Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T09:18:10.1542189Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-12-04T09:18:10.2097349Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:18:10.2818158Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-12-04T09:18:10.3326702Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-12-04T09:18:10.4684245Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-12-04T09:18:11.0981698Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-12-04T09:18:11.1026422Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-12-04T09:18:11.1061920Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-12-04T09:18:12.0551695Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-12-04T09:18:12.1543457Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-12-04T09:18:12.1571333Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:18:12.1574100Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:18:12.1577996Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:18:12.1582247Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:18:12.1586444Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:18:12.1590631Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:18:12.1594892Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:18:12.1599047Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:18:12.1637471Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-12-04T09:18:12.6203573Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-12-04T09:18:12.6205303Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-12-04T09:18:12.6207106Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-12-04T09:18:12.6208662Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-12-04T09:18:12.7204662Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-12-04T09:18:13.4158246Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-12-04T09:18:20.7167164Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-12-04T09:18:21.4605953Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-12-04T09:18:21.5113330Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-12-04T09:18:21.5336477Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-12-04T09:18:21.6739815Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-12-04T09:18:21.6931002Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-12-04T09:18:21.7139629Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-12-04T09:18:21.7368010Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-12-04T09:18:21.7390762Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:21.7393686Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:21.7427934Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T09:18:23.8268280Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T09:18:24.1171703Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-12-04T09:18:24.1744060Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T09:18:24.8990993Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-12-04T09:18:24.9154203Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-12-04T09:18:25.2534877Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-12-04T09:18:25.2565016Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:18:25.2568621Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-12-04T09:18:25.2603222Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-12-04T09:18:25.8167060Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-12-04T09:18:26.2411509Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-12-04T09:18:26.3275981Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-12-04T09:18:26.3419284Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-12-04T09:18:26.3591870Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-12-04T09:18:26.4149406Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-12-04T09:18:26.4518620Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-12-04T09:18:26.5065718Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-12-04T09:18:26.5442365Z Submodule path 'third_party/tensorpipe': checked out '2b4cd91092d335a697416b2a3cb398283246849d' 2025-12-04T09:18:26.5468627Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:18:26.5471898Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:18:26.5475442Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:18:26.5479415Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:18:26.5525951Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-12-04T09:18:27.8123958Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-12-04T09:18:27.8125223Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-12-04T09:18:27.8126463Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-12-04T09:18:27.8815747Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-12-04T09:18:27.9036322Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-12-04T09:18:27.9935472Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-12-04T09:18:28.0315105Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-12-04T09:18:28.0339707Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:18:28.0375031Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-12-04T09:18:28.2181954Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-12-04T09:18:28.2241586Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-12-04T09:18:28.2643659Z Entering 'android/libs/fbjni' 2025-12-04T09:18:28.2703052Z Entering 'third_party/FP16' 2025-12-04T09:18:28.2762731Z Entering 'third_party/FXdiv' 2025-12-04T09:18:28.2822279Z Entering 'third_party/NNPACK' 2025-12-04T09:18:28.2882407Z Entering 'third_party/NVTX' 2025-12-04T09:18:28.2942573Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:18:28.3001408Z Entering 'third_party/XNNPACK' 2025-12-04T09:18:28.3076331Z Entering 'third_party/aiter' 2025-12-04T09:18:28.3135967Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:18:28.3207811Z Entering 'third_party/benchmark' 2025-12-04T09:18:28.3267393Z Entering 'third_party/composable_kernel' 2025-12-04T09:18:28.3334961Z Entering 'third_party/cpp-httplib' 2025-12-04T09:18:28.3393734Z Entering 'third_party/cpuinfo' 2025-12-04T09:18:28.3452932Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:18:28.3511596Z Entering 'third_party/cutlass' 2025-12-04T09:18:28.3580410Z Entering 'third_party/fbgemm' 2025-12-04T09:18:28.3642005Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:18:28.3699295Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:18:28.3766664Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:18:28.3823854Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:18:28.3889628Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:18:28.3954649Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:18:28.4012478Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:18:28.4076346Z Entering 'third_party/flash-attention' 2025-12-04T09:18:28.4140723Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:18:28.4205432Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:18:28.4276274Z Entering 'third_party/flatbuffers' 2025-12-04T09:18:28.4338724Z Entering 'third_party/fmt' 2025-12-04T09:18:28.4398750Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:18:28.4462311Z Entering 'third_party/gloo' 2025-12-04T09:18:28.4523154Z Entering 'third_party/googletest' 2025-12-04T09:18:28.4583061Z Entering 'third_party/ideep' 2025-12-04T09:18:28.4640735Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:18:28.4711494Z Entering 'third_party/ittapi' 2025-12-04T09:18:28.4771970Z Entering 'third_party/kineto' 2025-12-04T09:18:28.4832425Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:18:28.4889694Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:18:28.4950672Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:18:28.5008142Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:18:28.5071238Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:18:28.5127369Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:18:28.5196421Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:18:28.5260634Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:18:28.5318422Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:18:28.5378131Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:18:28.5436979Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:18:28.5492553Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:28.5554513Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:28.5619795Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:18:28.5677410Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:18:28.5744860Z Entering 'third_party/kleidiai' 2025-12-04T09:18:28.5804029Z Entering 'third_party/mimalloc' 2025-12-04T09:18:28.5864908Z Entering 'third_party/nlohmann' 2025-12-04T09:18:28.5924602Z Entering 'third_party/onnx' 2025-12-04T09:18:28.6003362Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:18:28.6074757Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:18:28.6134903Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:18:28.6194490Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:18:28.6252322Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:18:28.6308705Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:18:28.6368711Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:18:28.6423994Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:18:28.6485432Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:18:28.6543136Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:28.6601122Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:28.6668010Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:18:28.6748791Z Entering 'third_party/pocketfft' 2025-12-04T09:18:28.6808370Z Entering 'third_party/protobuf' 2025-12-04T09:18:28.6869016Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:18:28.6927258Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:18:28.6989961Z Entering 'third_party/psimd' 2025-12-04T09:18:28.7049577Z Entering 'third_party/pthreadpool' 2025-12-04T09:18:28.7107907Z Entering 'third_party/pybind11' 2025-12-04T09:18:28.7168172Z Entering 'third_party/python-peachpy' 2025-12-04T09:18:28.7226566Z Entering 'third_party/sleef' 2025-12-04T09:18:28.7286950Z Entering 'third_party/tensorpipe' 2025-12-04T09:18:28.7344660Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:18:28.7403185Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:18:28.7459981Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:18:28.7517367Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:18:28.7573631Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:18:28.7660989Z ##[endgroup] 2025-12-04T09:18:28.7661456Z ##[group]Persisting credentials for submodules 2025-12-04T09:18:28.7667109Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-12-04T09:18:28.8061207Z Entering 'android/libs/fbjni' 2025-12-04T09:18:28.8139834Z Entering 'third_party/FP16' 2025-12-04T09:18:28.8216191Z Entering 'third_party/FXdiv' 2025-12-04T09:18:28.8294291Z Entering 'third_party/NNPACK' 2025-12-04T09:18:28.8371647Z Entering 'third_party/NVTX' 2025-12-04T09:18:28.8458076Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:18:28.8534330Z Entering 'third_party/XNNPACK' 2025-12-04T09:18:28.8629099Z Entering 'third_party/aiter' 2025-12-04T09:18:28.8711214Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:18:28.8799331Z Entering 'third_party/benchmark' 2025-12-04T09:18:28.8883473Z Entering 'third_party/composable_kernel' 2025-12-04T09:18:28.8971362Z Entering 'third_party/cpp-httplib' 2025-12-04T09:18:28.9049534Z Entering 'third_party/cpuinfo' 2025-12-04T09:18:28.9128381Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:18:28.9205979Z Entering 'third_party/cutlass' 2025-12-04T09:18:28.9292205Z Entering 'third_party/fbgemm' 2025-12-04T09:18:28.9373860Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:18:28.9450396Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:18:28.9535211Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:18:28.9611042Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:18:28.9699545Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:18:28.9774167Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:18:28.9854480Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:18:28.9936675Z Entering 'third_party/flash-attention' 2025-12-04T09:18:29.0013791Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:18:29.0098070Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:18:29.0186034Z Entering 'third_party/flatbuffers' 2025-12-04T09:18:29.0267397Z Entering 'third_party/fmt' 2025-12-04T09:18:29.0345491Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:18:29.0425587Z Entering 'third_party/gloo' 2025-12-04T09:18:29.0511913Z Entering 'third_party/googletest' 2025-12-04T09:18:29.0590480Z Entering 'third_party/ideep' 2025-12-04T09:18:29.0666356Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:18:29.0752047Z Entering 'third_party/ittapi' 2025-12-04T09:18:29.0827254Z Entering 'third_party/kineto' 2025-12-04T09:18:29.0904260Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:18:29.0979184Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:18:29.1056985Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:18:29.1134503Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:18:29.1212191Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:18:29.1287908Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:18:29.1372313Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:18:29.1453665Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:18:29.1531825Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:18:29.1611973Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:18:29.1691235Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:18:29.1767468Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:29.1847547Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:29.1934166Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:18:29.2013517Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:18:29.2094356Z Entering 'third_party/kleidiai' 2025-12-04T09:18:29.2173271Z Entering 'third_party/mimalloc' 2025-12-04T09:18:29.2252376Z Entering 'third_party/nlohmann' 2025-12-04T09:18:29.2331500Z Entering 'third_party/onnx' 2025-12-04T09:18:29.2429320Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:18:29.2511783Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:18:29.2593027Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:18:29.2669415Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:18:29.2746285Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:18:29.2821460Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:18:29.2900016Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:18:29.2974599Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:18:29.3049937Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:18:29.3123565Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:29.3202827Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:29.3285484Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:18:29.3386522Z Entering 'third_party/pocketfft' 2025-12-04T09:18:29.3464228Z Entering 'third_party/protobuf' 2025-12-04T09:18:29.3550212Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:18:29.3626717Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:18:29.3709053Z Entering 'third_party/psimd' 2025-12-04T09:18:29.3789015Z Entering 'third_party/pthreadpool' 2025-12-04T09:18:29.3866651Z Entering 'third_party/pybind11' 2025-12-04T09:18:29.3947042Z Entering 'third_party/python-peachpy' 2025-12-04T09:18:29.4025086Z Entering 'third_party/sleef' 2025-12-04T09:18:29.4103707Z Entering 'third_party/tensorpipe' 2025-12-04T09:18:29.4181659Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:18:29.4259151Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:18:29.4333062Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:18:29.4414662Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:18:29.4491025Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:18:29.4601352Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-12-04T09:18:29.5008644Z Entering 'android/libs/fbjni' 2025-12-04T09:18:29.5084104Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T09:18:29.5108877Z Entering 'third_party/FP16' 2025-12-04T09:18:29.5181556Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T09:18:29.5206297Z Entering 'third_party/FXdiv' 2025-12-04T09:18:29.5283957Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T09:18:29.5309464Z Entering 'third_party/NNPACK' 2025-12-04T09:18:29.5382920Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T09:18:29.5407646Z Entering 'third_party/NVTX' 2025-12-04T09:18:29.5480640Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T09:18:29.5507216Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:18:29.5578453Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T09:18:29.5603414Z Entering 'third_party/XNNPACK' 2025-12-04T09:18:29.5679587Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T09:18:29.5719973Z Entering 'third_party/aiter' 2025-12-04T09:18:29.5792850Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T09:18:29.5817263Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:18:29.5891102Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T09:18:29.5926721Z Entering 'third_party/benchmark' 2025-12-04T09:18:29.6005527Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:18:29.6031048Z Entering 'third_party/composable_kernel' 2025-12-04T09:18:29.6103064Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T09:18:29.6137847Z Entering 'third_party/cpp-httplib' 2025-12-04T09:18:29.6208023Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T09:18:29.6234291Z Entering 'third_party/cpuinfo' 2025-12-04T09:18:29.6307628Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T09:18:29.6333083Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:18:29.6403892Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T09:18:29.6433512Z Entering 'third_party/cutlass' 2025-12-04T09:18:29.6510123Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T09:18:29.6546298Z Entering 'third_party/fbgemm' 2025-12-04T09:18:29.6616439Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T09:18:29.6643942Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:18:29.6714783Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T09:18:29.6739031Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:18:29.6810095Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T09:18:29.6843874Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:18:29.6913872Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T09:18:29.6938298Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:18:29.7010061Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T09:18:29.7043707Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:18:29.7113993Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T09:18:29.7142958Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:18:29.7214608Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T09:18:29.7238437Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:18:29.7309835Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T09:18:29.7338164Z Entering 'third_party/flash-attention' 2025-12-04T09:18:29.7411397Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T09:18:29.7436042Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:18:29.7505301Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T09:18:29.7536626Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:18:29.7610069Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T09:18:29.7645137Z Entering 'third_party/flatbuffers' 2025-12-04T09:18:29.7725103Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T09:18:29.7754555Z Entering 'third_party/fmt' 2025-12-04T09:18:29.7824899Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:18:29.7850525Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:18:29.7927004Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T09:18:29.7952929Z Entering 'third_party/gloo' 2025-12-04T09:18:29.8023019Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T09:18:29.8049486Z Entering 'third_party/googletest' 2025-12-04T09:18:29.8125617Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:18:29.8150897Z Entering 'third_party/ideep' 2025-12-04T09:18:29.8222798Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T09:18:29.8246003Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:18:29.8316485Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T09:18:29.8349912Z Entering 'third_party/ittapi' 2025-12-04T09:18:29.8424763Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T09:18:29.8450311Z Entering 'third_party/kineto' 2025-12-04T09:18:29.8525002Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T09:18:29.8550465Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:18:29.8621038Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T09:18:29.8643340Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:18:29.8715475Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T09:18:29.8741997Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:18:29.8814415Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T09:18:29.8837647Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:18:29.8908159Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:18:29.8932259Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:18:29.9004716Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T09:18:29.9027329Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:18:29.9101243Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T09:18:29.9128988Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:18:29.9200696Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T09:18:29.9226543Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:18:29.9307451Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:18:29.9331901Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:18:29.9403483Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T09:18:29.9428865Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:18:29.9504984Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T09:18:29.9529403Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:18:29.9601267Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:18:29.9623544Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:29.9696795Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:18:29.9724316Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:29.9797558Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:18:29.9830691Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:18:29.9906101Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T09:18:29.9929862Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:18:30.0001071Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T09:18:30.0028243Z Entering 'third_party/kleidiai' 2025-12-04T09:18:30.0101434Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T09:18:30.0127956Z Entering 'third_party/mimalloc' 2025-12-04T09:18:30.0206724Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T09:18:30.0232293Z Entering 'third_party/nlohmann' 2025-12-04T09:18:30.0303287Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T09:18:30.0330008Z Entering 'third_party/onnx' 2025-12-04T09:18:30.0401742Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T09:18:30.0442037Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:18:30.0516188Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:18:30.0546808Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:18:30.0618797Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T09:18:30.0645040Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:18:30.0714261Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:18:30.0743013Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:18:30.0812621Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:18:30.0837843Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:18:30.0910499Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T09:18:30.0934567Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:18:30.1006380Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T09:18:30.1031008Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:18:30.1102383Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T09:18:30.1127334Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:18:30.1199724Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T09:18:30.1223220Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:18:30.1294982Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:18:30.1315838Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:30.1391073Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:18:30.1419314Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:30.1492386Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:18:30.1520856Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:18:30.1591950Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T09:18:30.1639158Z Entering 'third_party/pocketfft' 2025-12-04T09:18:30.1710712Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T09:18:30.1736480Z Entering 'third_party/protobuf' 2025-12-04T09:18:30.1810694Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T09:18:30.1838361Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:18:30.1913330Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:18:30.1937247Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:18:30.2015873Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:18:30.2047001Z Entering 'third_party/psimd' 2025-12-04T09:18:30.2122153Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T09:18:30.2146326Z Entering 'third_party/pthreadpool' 2025-12-04T09:18:30.2218222Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T09:18:30.2247093Z Entering 'third_party/pybind11' 2025-12-04T09:18:30.2326119Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:18:30.2355193Z Entering 'third_party/python-peachpy' 2025-12-04T09:18:30.2424957Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T09:18:30.2450111Z Entering 'third_party/sleef' 2025-12-04T09:18:30.2526758Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T09:18:30.2553149Z Entering 'third_party/tensorpipe' 2025-12-04T09:18:30.2631662Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T09:18:30.2656702Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:18:30.2730066Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:18:30.2755390Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:18:30.2825299Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T09:18:30.2849813Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:18:30.2931298Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T09:18:30.2956078Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:18:30.3029869Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:18:30.3051717Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:18:30.3139916Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T09:18:30.6005953Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-12-04T09:18:30.6414759Z Entering 'android/libs/fbjni' 2025-12-04T09:18:30.6476812Z Entering 'third_party/FP16' 2025-12-04T09:18:30.6543091Z Entering 'third_party/FXdiv' 2025-12-04T09:18:30.6604986Z Entering 'third_party/NNPACK' 2025-12-04T09:18:30.6666434Z Entering 'third_party/NVTX' 2025-12-04T09:18:30.6727090Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:18:30.6786850Z Entering 'third_party/XNNPACK' 2025-12-04T09:18:30.6861551Z Entering 'third_party/aiter' 2025-12-04T09:18:30.6921225Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:18:30.6989827Z Entering 'third_party/benchmark' 2025-12-04T09:18:30.7051739Z Entering 'third_party/composable_kernel' 2025-12-04T09:18:30.7122224Z Entering 'third_party/cpp-httplib' 2025-12-04T09:18:30.7181913Z Entering 'third_party/cpuinfo' 2025-12-04T09:18:30.7244989Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:18:30.7304942Z Entering 'third_party/cutlass' 2025-12-04T09:18:30.7376433Z Entering 'third_party/fbgemm' 2025-12-04T09:18:30.7438399Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:18:30.7496321Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:18:30.7565314Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:18:30.7623068Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:18:30.7689187Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:18:30.7746412Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:18:30.7802486Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:18:30.7865655Z Entering 'third_party/flash-attention' 2025-12-04T09:18:30.7925005Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:18:30.7990764Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:18:30.8061030Z Entering 'third_party/flatbuffers' 2025-12-04T09:18:30.8123817Z Entering 'third_party/fmt' 2025-12-04T09:18:30.8183931Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:18:30.8245500Z Entering 'third_party/gloo' 2025-12-04T09:18:30.8304794Z Entering 'third_party/googletest' 2025-12-04T09:18:30.8364339Z Entering 'third_party/ideep' 2025-12-04T09:18:30.8423755Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:18:30.8492133Z Entering 'third_party/ittapi' 2025-12-04T09:18:30.8553690Z Entering 'third_party/kineto' 2025-12-04T09:18:30.8611596Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:18:30.8672231Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:18:30.8732709Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:18:30.8792104Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:18:30.8856232Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:18:30.8913486Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:18:30.8976058Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:18:30.9033844Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:18:30.9097122Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:18:30.9155835Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:18:30.9216052Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:18:30.9272414Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:30.9334435Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:30.9408012Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:18:30.9464996Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:18:30.9525725Z Entering 'third_party/kleidiai' 2025-12-04T09:18:30.9591851Z Entering 'third_party/mimalloc' 2025-12-04T09:18:30.9654973Z Entering 'third_party/nlohmann' 2025-12-04T09:18:30.9715646Z Entering 'third_party/onnx' 2025-12-04T09:18:30.9791856Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:18:30.9858255Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:18:30.9917025Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:18:30.9978975Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:18:31.0038974Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:18:31.0094639Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:18:31.0153794Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:18:31.0213462Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:18:31.0271577Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:18:31.0327083Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:31.0386703Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:31.0448350Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:18:31.0527811Z Entering 'third_party/pocketfft' 2025-12-04T09:18:31.0588837Z Entering 'third_party/protobuf' 2025-12-04T09:18:31.0657790Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:18:31.0714192Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:18:31.0775802Z Entering 'third_party/psimd' 2025-12-04T09:18:31.0835061Z Entering 'third_party/pthreadpool' 2025-12-04T09:18:31.0895838Z Entering 'third_party/pybind11' 2025-12-04T09:18:31.0954539Z Entering 'third_party/python-peachpy' 2025-12-04T09:18:31.1013550Z Entering 'third_party/sleef' 2025-12-04T09:18:31.1074481Z Entering 'third_party/tensorpipe' 2025-12-04T09:18:31.1135890Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:18:31.1191944Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:18:31.1253061Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:18:31.1310145Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:18:31.1365543Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:18:31.1456696Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-12-04T09:18:31.1856324Z Entering 'android/libs/fbjni' 2025-12-04T09:18:31.1916679Z Entering 'third_party/FP16' 2025-12-04T09:18:31.1983426Z Entering 'third_party/FXdiv' 2025-12-04T09:18:31.2043087Z Entering 'third_party/NNPACK' 2025-12-04T09:18:31.2103602Z Entering 'third_party/NVTX' 2025-12-04T09:18:31.2163683Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:18:31.2231464Z Entering 'third_party/XNNPACK' 2025-12-04T09:18:31.2305728Z Entering 'third_party/aiter' 2025-12-04T09:18:31.2371260Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:18:31.2439756Z Entering 'third_party/benchmark' 2025-12-04T09:18:31.2501465Z Entering 'third_party/composable_kernel' 2025-12-04T09:18:31.2574040Z Entering 'third_party/cpp-httplib' 2025-12-04T09:18:31.2637364Z Entering 'third_party/cpuinfo' 2025-12-04T09:18:31.2697402Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:18:31.2756794Z Entering 'third_party/cutlass' 2025-12-04T09:18:31.2830261Z Entering 'third_party/fbgemm' 2025-12-04T09:18:31.2900934Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:18:31.2958307Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:18:31.3023085Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:18:31.3079890Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:18:31.3146012Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:18:31.3202758Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:18:31.3259626Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:18:31.3328070Z Entering 'third_party/flash-attention' 2025-12-04T09:18:31.3386917Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:18:31.3450838Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:18:31.3518421Z Entering 'third_party/flatbuffers' 2025-12-04T09:18:31.3584530Z Entering 'third_party/fmt' 2025-12-04T09:18:31.3644364Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:18:31.3704425Z Entering 'third_party/gloo' 2025-12-04T09:18:31.3765260Z Entering 'third_party/googletest' 2025-12-04T09:18:31.3829720Z Entering 'third_party/ideep' 2025-12-04T09:18:31.3887631Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:18:31.3955254Z Entering 'third_party/ittapi' 2025-12-04T09:18:31.4014710Z Entering 'third_party/kineto' 2025-12-04T09:18:31.4074700Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:18:31.4130342Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:18:31.4188833Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:18:31.4252228Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:18:31.4310824Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:18:31.4369412Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:18:31.4431735Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:18:31.4492894Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:18:31.4555905Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:18:31.4615418Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:18:31.4674878Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:18:31.4734103Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:31.4794983Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:31.4860824Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:18:31.4918891Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:18:31.4980223Z Entering 'third_party/kleidiai' 2025-12-04T09:18:31.5041142Z Entering 'third_party/mimalloc' 2025-12-04T09:18:31.5103697Z Entering 'third_party/nlohmann' 2025-12-04T09:18:31.5166104Z Entering 'third_party/onnx' 2025-12-04T09:18:31.5242637Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:18:31.5306014Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:18:31.5370926Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:18:31.5430333Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:18:31.5486971Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:18:31.5543202Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:18:31.5601426Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:18:31.5657714Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:18:31.5713275Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:18:31.5768540Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:31.5827292Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:31.5893608Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:18:31.5973615Z Entering 'third_party/pocketfft' 2025-12-04T09:18:31.6032773Z Entering 'third_party/protobuf' 2025-12-04T09:18:31.6094747Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:18:31.6153002Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:18:31.6215231Z Entering 'third_party/psimd' 2025-12-04T09:18:31.6276165Z Entering 'third_party/pthreadpool' 2025-12-04T09:18:31.6343843Z Entering 'third_party/pybind11' 2025-12-04T09:18:31.6405322Z Entering 'third_party/python-peachpy' 2025-12-04T09:18:31.6468135Z Entering 'third_party/sleef' 2025-12-04T09:18:31.6528002Z Entering 'third_party/tensorpipe' 2025-12-04T09:18:31.6588399Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:18:31.6644980Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:18:31.6704997Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:18:31.6762161Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:18:31.6821101Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:18:31.6909237Z ##[endgroup] 2025-12-04T09:18:31.6953417Z [command]/usr/bin/git log -1 --format=%H 2025-12-04T09:18:31.6981611Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:18:31.7113818Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-12-04T09:18:31.7114168Z cd "${GITHUB_WORKSPACE}" 2025-12-04T09:18:31.7114471Z # Clean stale submodule dirs 2025-12-04T09:18:31.7114778Z if [ -z "${NO_SUDO}" ]; then 2025-12-04T09:18:31.7115155Z  sudo git submodule foreach --recursive git clean -ffdx 2025-12-04T09:18:31.7115522Z else 2025-12-04T09:18:31.7115822Z  git submodule foreach --recursive git clean -ffdx 2025-12-04T09:18:31.7116164Z fi 2025-12-04T09:18:31.7127727Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:31.7128081Z env: 2025-12-04T09:18:31.7128292Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:31.7128537Z NO_SUDO: true 2025-12-04T09:18:31.7128755Z ##[endgroup] 2025-12-04T09:18:31.7553110Z Entering 'android/libs/fbjni' 2025-12-04T09:18:31.7600554Z Entering 'third_party/FP16' 2025-12-04T09:18:31.7647172Z Entering 'third_party/FXdiv' 2025-12-04T09:18:31.7694078Z Entering 'third_party/NNPACK' 2025-12-04T09:18:31.7744892Z Entering 'third_party/NVTX' 2025-12-04T09:18:31.7798382Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:18:31.7847641Z Entering 'third_party/XNNPACK' 2025-12-04T09:18:31.8005369Z Entering 'third_party/aiter' 2025-12-04T09:18:31.8065627Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:18:31.8213985Z Entering 'third_party/benchmark' 2025-12-04T09:18:31.8264126Z Entering 'third_party/composable_kernel' 2025-12-04T09:18:31.8426194Z Entering 'third_party/cpp-httplib' 2025-12-04T09:18:31.8475686Z Entering 'third_party/cpuinfo' 2025-12-04T09:18:31.8529059Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:18:31.8579891Z Entering 'third_party/cutlass' 2025-12-04T09:18:31.8714797Z Entering 'third_party/fbgemm' 2025-12-04T09:18:31.8799111Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:18:31.8844069Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:18:31.9001485Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:18:31.9050569Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:18:31.9182055Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:18:31.9228634Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:18:31.9272626Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:18:31.9338701Z Entering 'third_party/flash-attention' 2025-12-04T09:18:31.9394575Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:18:31.9531623Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:18:31.9655943Z Entering 'third_party/flatbuffers' 2025-12-04T09:18:31.9758474Z Entering 'third_party/fmt' 2025-12-04T09:18:31.9805555Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:18:31.9853530Z Entering 'third_party/gloo' 2025-12-04T09:18:31.9902111Z Entering 'third_party/googletest' 2025-12-04T09:18:31.9951554Z Entering 'third_party/ideep' 2025-12-04T09:18:31.9993720Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:18:32.0111290Z Entering 'third_party/ittapi' 2025-12-04T09:18:32.0160249Z Entering 'third_party/kineto' 2025-12-04T09:18:32.0209687Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:18:32.0262708Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:18:32.0326355Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:18:32.0371779Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:18:32.0422628Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:18:32.0463281Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:18:32.0509121Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:18:32.0553264Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:18:32.0600253Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:18:32.0659148Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:18:32.0703607Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:18:32.0747852Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:32.0815489Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:32.0876895Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:18:32.0921075Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:18:32.0970950Z Entering 'third_party/kleidiai' 2025-12-04T09:18:32.1027702Z Entering 'third_party/mimalloc' 2025-12-04T09:18:32.1078044Z Entering 'third_party/nlohmann' 2025-12-04T09:18:32.1143094Z Entering 'third_party/onnx' 2025-12-04T09:18:32.1616542Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:18:32.1671517Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:18:32.1750663Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:18:32.1794366Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:18:32.1841543Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:18:32.1885521Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:18:32.1956675Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:18:32.1991550Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:18:32.2037341Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:18:32.2079775Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:18:32.2150738Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:18:32.2200445Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:18:32.2562492Z Entering 'third_party/pocketfft' 2025-12-04T09:18:32.2611058Z Entering 'third_party/protobuf' 2025-12-04T09:18:32.2724532Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:18:32.2771696Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:18:32.2823857Z Entering 'third_party/psimd' 2025-12-04T09:18:32.2869914Z Entering 'third_party/pthreadpool' 2025-12-04T09:18:32.2917896Z Entering 'third_party/pybind11' 2025-12-04T09:18:32.2968610Z Entering 'third_party/python-peachpy' 2025-12-04T09:18:32.3016429Z Entering 'third_party/sleef' 2025-12-04T09:18:32.3066606Z Entering 'third_party/tensorpipe' 2025-12-04T09:18:32.3114806Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:18:32.3161210Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:18:32.3203463Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:18:32.3252243Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:18:32.3293566Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:18:32.3467779Z Prepare all required actions 2025-12-04T09:18:32.3468311Z Getting action download info 2025-12-04T09:18:32.5050777Z ##[group]Run ./.github/actions/setup-linux 2025-12-04T09:18:32.5051087Z env: 2025-12-04T09:18:32.5051311Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:32.5051573Z ##[endgroup] 2025-12-04T09:18:32.5088199Z ##[group]Run set -euo pipefail 2025-12-04T09:18:32.5088522Z set -euo pipefail 2025-12-04T09:18:32.5088812Z function get_ec2_metadata() { 2025-12-04T09:18:32.5089186Z  # Pulled from instance metadata endpoint for EC2 2025-12-04T09:18:32.5089780Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-12-04T09:18:32.5090325Z  category=$1 2025-12-04T09:18:32.5090845Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-12-04T09:18:32.5091269Z  runner_name_str=i-016559b86632fc7da 2025-12-04T09:18:32.5091624Z  if [[ -f /.inarc ]]; then 2025-12-04T09:18:32.5091963Z  echo "ARC Runner, no info on ec2 metadata" 2025-12-04T09:18:32.5092342Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-12-04T09:18:32.5092788Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-12-04T09:18:32.5093182Z  else 2025-12-04T09:18:32.5093976Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-12-04T09:18:32.5094804Z  fi 2025-12-04T09:18:32.5095015Z } 2025-12-04T09:18:32.5095265Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-12-04T09:18:32.5095690Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-12-04T09:18:32.5096172Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-12-04T09:18:32.5096606Z echo "system info $(uname -a)" 2025-12-04T09:18:32.5105769Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:32.5106129Z env: 2025-12-04T09:18:32.5106339Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:32.5106596Z ##[endgroup] 2025-12-04T09:18:32.5284254Z ami-id: ami-08982f1c5bf93d976 2025-12-04T09:18:32.5406980Z instance-id: i-016559b86632fc7da 2025-12-04T09:18:32.5531143Z instance-type: g5.4xlarge 2025-12-04T09:18:32.5546740Z system info Linux ip-10-0-66-49.ec2.internal 6.1.150-174.273.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Sep 9 12:21:26 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-12-04T09:18:32.5581965Z ##[group]Run if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T09:18:32.5582425Z if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T09:18:32.5591806Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:32.5592164Z env: 2025-12-04T09:18:32.5592369Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:32.5592622Z ##[endgroup] 2025-12-04T09:18:34.1724894Z Thu Dec 4 09:18:34 2025 2025-12-04T09:18:34.1725316Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:18:34.1725828Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:18:34.1726326Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:18:34.1726876Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:18:34.1727513Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:18:34.1727963Z | | | MIG M. | 2025-12-04T09:18:34.1728317Z |=========================================+========================+======================| 2025-12-04T09:18:34.1818516Z | 0 NVIDIA A10G Off | 00000000:00:1E.0 Off | 0 | 2025-12-04T09:18:34.1819333Z | 0% 25C P0 56W / 300W | 0MiB / 23028MiB | 3% Default | 2025-12-04T09:18:34.1819729Z | | | N/A | 2025-12-04T09:18:34.1820134Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:18:34.1820418Z 2025-12-04T09:18:34.1821130Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:18:34.1821638Z | Processes: | 2025-12-04T09:18:34.1822104Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:18:34.1822533Z | ID ID Usage | 2025-12-04T09:18:34.1823162Z |=========================================================================================| 2025-12-04T09:18:34.1823831Z | No running processes found | 2025-12-04T09:18:34.1824312Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:18:34.6120744Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:18:34.6121631Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:18:34.6134695Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:34.6135051Z env: 2025-12-04T09:18:34.6135753Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:34.6136011Z ##[endgroup] 2025-12-04T09:18:34.6195607Z ##[group]Run if systemctl is-active --quiet docker; then 2025-12-04T09:18:34.6196052Z if systemctl is-active --quiet docker; then 2025-12-04T09:18:34.6196425Z  echo "Docker daemon is running..."; 2025-12-04T09:18:34.6196752Z else 2025-12-04T09:18:34.6197094Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-12-04T09:18:34.6197504Z fi 2025-12-04T09:18:34.6207228Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:34.6207587Z env: 2025-12-04T09:18:34.6207790Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:34.6208044Z ##[endgroup] 2025-12-04T09:18:34.6305220Z Docker daemon is running... 2025-12-04T09:18:34.6346551Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:18:34.6346836Z with: 2025-12-04T09:18:34.6347037Z shell: bash 2025-12-04T09:18:34.6347252Z timeout_minutes: 5 2025-12-04T09:18:34.6347484Z max_attempts: 3 2025-12-04T09:18:34.6347709Z retry_wait_seconds: 30 2025-12-04T09:18:34.6349844Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-12-04T09:18:34.6351989Z polling_interval_seconds: 1 2025-12-04T09:18:34.6352266Z warning_on_retry: true 2025-12-04T09:18:34.6352525Z continue_on_error: false 2025-12-04T09:18:34.6352764Z env: 2025-12-04T09:18:34.6352971Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:34.6353229Z AWS_RETRY_MODE: standard 2025-12-04T09:18:34.6353477Z AWS_MAX_ATTEMPTS: 5 2025-12-04T09:18:34.6353729Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:18:34.6353996Z ##[endgroup] 2025-12-04T09:18:35.8142418Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:18:35.8143397Z Configure a credential helper to remove this warning. See 2025-12-04T09:18:35.8143996Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:18:35.8144373Z 2025-12-04T09:18:35.8144482Z Login Succeeded 2025-12-04T09:18:36.7203241Z Command completed after 1 attempt(s). 2025-12-04T09:18:36.7278914Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:18:36.7279411Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:18:36.7279853Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:18:36.7291010Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:36.7291364Z env: 2025-12-04T09:18:36.7291577Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:36.7291827Z ##[endgroup] 2025-12-04T09:18:36.7419030Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:18:36.7419576Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:18:36.7419997Z # shellcheck disable=SC2046 2025-12-04T09:18:36.7420328Z docker stop $(docker ps -q) || true 2025-12-04T09:18:36.7420663Z # Prune all of the docker images 2025-12-04T09:18:36.7420985Z docker system prune -af 2025-12-04T09:18:36.7431019Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:36.7431370Z env: 2025-12-04T09:18:36.7431581Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:36.7431838Z ##[endgroup] 2025-12-04T09:18:36.7741181Z "docker stop" requires at least 1 argument. 2025-12-04T09:18:36.7741583Z See 'docker stop --help'. 2025-12-04T09:18:36.7741757Z 2025-12-04T09:18:36.7741923Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-12-04T09:18:36.7742190Z 2025-12-04T09:18:36.7742300Z Stop one or more running containers 2025-12-04T09:18:36.7964830Z Total reclaimed space: 0B 2025-12-04T09:18:36.8147424Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-12-04T09:18:36.8147881Z with: 2025-12-04T09:18:36.8148640Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:36.8149481Z use-custom-docker-registry: true 2025-12-04T09:18:36.8149792Z docker-build-dir: .ci/docker 2025-12-04T09:18:36.8150085Z docker-build-script: ./build.sh 2025-12-04T09:18:36.8150371Z working-directory: . 2025-12-04T09:18:36.8150709Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:36.8151094Z force-push: false 2025-12-04T09:18:36.8151311Z env: 2025-12-04T09:18:36.8151517Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:36.8151771Z ##[endgroup] 2025-12-04T09:18:36.8170404Z ##[group]Run set -ex 2025-12-04T09:18:36.8170680Z set -ex 2025-12-04T09:18:36.8170913Z  2025-12-04T09:18:36.8171323Z # If the docker build directory or the build script doesn't exist, the action will 2025-12-04T09:18:36.8171960Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-12-04T09:18:36.8172513Z # job could then download the pre-built image as usual 2025-12-04T09:18:36.8173165Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-12-04T09:18:36.8173770Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8174089Z else 2025-12-04T09:18:36.8174351Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8174788Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8175169Z  2025-12-04T09:18:36.8175699Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-12-04T09:18:36.8176296Z  exit 0 2025-12-04T09:18:36.8176504Z fi 2025-12-04T09:18:36.8176709Z  2025-12-04T09:18:36.8177038Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-12-04T09:18:36.8177610Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-12-04T09:18:36.8178115Z  # use it as it is, but first let's extract the tag 2025-12-04T09:18:36.8178581Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-12-04T09:18:36.8179074Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8179536Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8179914Z else 2025-12-04T09:18:36.8180180Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-12-04T09:18:36.8180552Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-12-04T09:18:36.8181108Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-12-04T09:18:36.8181426Z  fi 2025-12-04T09:18:36.8181857Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-12-04T09:18:36.8182426Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8183035Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8183697Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8184108Z fi 2025-12-04T09:18:36.8193197Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:36.8193552Z env: 2025-12-04T09:18:36.8193759Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:36.8194009Z REPO_NAME: pytorch 2025-12-04T09:18:36.8194914Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:36.8195745Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T09:18:36.8196036Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-12-04T09:18:36.8196399Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:36.8196796Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-12-04T09:18:36.8197084Z CUSTOM_TAG_PREFIX: 2025-12-04T09:18:36.8197322Z ##[endgroup] 2025-12-04T09:18:36.8228139Z + [[ -d .ci/docker ]] 2025-12-04T09:18:36.8228413Z + [[ -f .ci/docker/./build.sh ]] 2025-12-04T09:18:36.8228688Z + [[ true == \t\r\u\e ]] 2025-12-04T09:18:36.8228935Z + echo skip=false 2025-12-04T09:18:36.8229930Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-12-04T09:18:36.8237135Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:36.8237937Z ++ awk -F '[:,]' '{print $2}' 2025-12-04T09:18:36.8264561Z + DOCKER_TAG=pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:36.8265419Z + echo docker-tag=pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:36.8266516Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:36.8303876Z ##[group]Run set +e 2025-12-04T09:18:36.8304158Z set +e 2025-12-04T09:18:36.8304382Z set -x 2025-12-04T09:18:36.8304601Z  2025-12-04T09:18:36.8304799Z login() { 2025-12-04T09:18:36.8305258Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T09:18:36.8305755Z } 2025-12-04T09:18:36.8305957Z  2025-12-04T09:18:36.8306168Z retry () { 2025-12-04T09:18:36.8306436Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T09:18:36.8306734Z } 2025-12-04T09:18:36.8306926Z  2025-12-04T09:18:36.8307152Z retry login "${DOCKER_REGISTRY}" 2025-12-04T09:18:36.8307445Z  2025-12-04T09:18:36.8307654Z START_TIME=$(date +%s) 2025-12-04T09:18:36.8307942Z # Wait up to 120 minutes 2025-12-04T09:18:36.8308294Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-12-04T09:18:36.8308757Z  # Check if image already exists, if it does then skip building it 2025-12-04T09:18:36.8309229Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-12-04T09:18:36.8309570Z  exit 0 2025-12-04T09:18:36.8309798Z  fi 2025-12-04T09:18:36.8309998Z  2025-12-04T09:18:36.8310546Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-12-04T09:18:36.8311174Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-12-04T09:18:36.8311791Z  # latter, it will wait for the Docker images to become available before continuing 2025-12-04T09:18:36.8312284Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-12-04T09:18:36.8312663Z  # It's a Docker build job, let's build the image 2025-12-04T09:18:36.8312989Z  break 2025-12-04T09:18:36.8313201Z  else 2025-12-04T09:18:36.8313532Z  # It's a regular build job, wait for the image to become available 2025-12-04T09:18:36.8313927Z  sleep 300 2025-12-04T09:18:36.8314158Z  fi 2025-12-04T09:18:36.8314363Z done 2025-12-04T09:18:36.8314566Z  2025-12-04T09:18:36.8314895Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-12-04T09:18:36.8315590Z # be empty. The default action would be to continue rebuild the image 2025-12-04T09:18:36.8316085Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-12-04T09:18:36.8316520Z  # if we're on the base branch then use the parent commit 2025-12-04T09:18:36.8316901Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-12-04T09:18:36.8317206Z else 2025-12-04T09:18:36.8317522Z  # otherwise we're on a PR, so use the most recent base commit 2025-12-04T09:18:36.8317981Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-12-04T09:18:36.8318311Z fi 2025-12-04T09:18:36.8318510Z  2025-12-04T09:18:36.8318733Z if [[ -z "${MERGE_BASE}" ]]; then 2025-12-04T09:18:36.8319071Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8319385Z  2025-12-04T09:18:36.8319824Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-12-04T09:18:36.8320345Z  exit 0 2025-12-04T09:18:36.8320561Z fi 2025-12-04T09:18:36.8320765Z  2025-12-04T09:18:36.8321066Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-12-04T09:18:36.8321704Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-12-04T09:18:36.8322257Z  exit 1 2025-12-04T09:18:36.8322476Z fi 2025-12-04T09:18:36.8322669Z  2025-12-04T09:18:36.8323025Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-12-04T09:18:36.8323656Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-12-04T09:18:36.8324218Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-12-04T09:18:36.8324857Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-12-04T09:18:36.8325587Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-12-04T09:18:36.8326017Z fi 2025-12-04T09:18:36.8326225Z  2025-12-04T09:18:36.8326476Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:18:36.8335209Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:36.8335867Z env: 2025-12-04T09:18:36.8336078Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:36.8336340Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T09:18:36.8336683Z BASE_REVISION: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:18:36.8337547Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:36.8338613Z DOCKER_TAG: pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:36.8339401Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:36.8339771Z DOCKER_PUSH: 2025-12-04T09:18:36.8339990Z ##[endgroup] 2025-12-04T09:18:36.8373910Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:36.8374330Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:36.8377148Z + aws ecr get-login-password --region us-east-1 2025-12-04T09:18:36.8379099Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:37.3486187Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:18:37.3486784Z Configure a credential helper to remove this warning. See 2025-12-04T09:18:37.3487397Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:18:37.3487832Z 2025-12-04T09:18:37.3488222Z Login Succeeded 2025-12-04T09:18:37.3514228Z ++ date +%s 2025-12-04T09:18:37.3527546Z + START_TIME=1764839917 2025-12-04T09:18:37.3532198Z ++ date +%s 2025-12-04T09:18:37.3546227Z + [[ 1764832717 -lt 1764839917 ]] 2025-12-04T09:18:37.3547095Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:37.5864790Z { 2025-12-04T09:18:37.5865073Z "schemaVersion": 2, 2025-12-04T09:18:37.5865571Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-12-04T09:18:37.5866067Z "config": { 2025-12-04T09:18:37.5866400Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-12-04T09:18:37.5866788Z "size": 34864, 2025-12-04T09:18:37.5867175Z "digest": "sha256:add7313791033822205cdb3cf32096534b2cfaa4855bd48119b59000bfe00301" 2025-12-04T09:18:37.5867620Z }, 2025-12-04T09:18:37.5867850Z "layers": [ 2025-12-04T09:18:37.5868052Z { 2025-12-04T09:18:37.5868372Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5868761Z "size": 30447951, 2025-12-04T09:18:37.5869187Z "digest": "sha256:63e5bc7682b85ae57a1221210f64d62e7a90b0a30f19af4ca734b8242ae49d63" 2025-12-04T09:18:37.5869643Z }, 2025-12-04T09:18:37.5869824Z { 2025-12-04T09:18:37.5870138Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5870532Z "size": 1554, 2025-12-04T09:18:37.5870911Z "digest": "sha256:0678d56345c994444b77bb70b1177189d23e794748b1d75ffc45d227c7dea94a" 2025-12-04T09:18:37.5871341Z }, 2025-12-04T09:18:37.5871522Z { 2025-12-04T09:18:37.5871850Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5872274Z "size": 313275661, 2025-12-04T09:18:37.5872708Z "digest": "sha256:45f5c9ddfce78349dff3d5edfbaa0310ae17311f66abdcd7e00fa21b500e801c" 2025-12-04T09:18:37.5873170Z }, 2025-12-04T09:18:37.5873370Z { 2025-12-04T09:18:37.5873696Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5874089Z "size": 787, 2025-12-04T09:18:37.5874479Z "digest": "sha256:086b1df51ac1162d9c45698e9dfaf91c6c222c8bd9ab01797ac8f9344bc8044f" 2025-12-04T09:18:37.5874932Z }, 2025-12-04T09:18:37.5875115Z { 2025-12-04T09:18:37.5875434Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5875824Z "size": 106, 2025-12-04T09:18:37.5876235Z "digest": "sha256:fe8a7b64bf98352f89057bcba66beef2fb44cc05fbd3606abccd8e86cf476234" 2025-12-04T09:18:37.5876687Z }, 2025-12-04T09:18:37.5876865Z { 2025-12-04T09:18:37.5877183Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5877586Z "size": 703, 2025-12-04T09:18:37.5878015Z "digest": "sha256:7680723e9a578033dd106b45784c639f06cc8adb1f5239ec513d9de01087c1af" 2025-12-04T09:18:37.5878452Z }, 2025-12-04T09:18:37.5878647Z { 2025-12-04T09:18:37.5878963Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5879365Z "size": 1216, 2025-12-04T09:18:37.5879765Z "digest": "sha256:9c5027aeeb4e3101f48c1d2e400c387110e1009e42497ee801f1b4b7f7edb5c0" 2025-12-04T09:18:37.5880444Z }, 2025-12-04T09:18:37.5880632Z { 2025-12-04T09:18:37.5880954Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5881352Z "size": 483, 2025-12-04T09:18:37.5881728Z "digest": "sha256:9a56521103600bd37a1e7c1191b5136c2d738c092f8a6701499f7068a32c2628" 2025-12-04T09:18:37.5882158Z }, 2025-12-04T09:18:37.5882356Z { 2025-12-04T09:18:37.5882706Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5883108Z "size": 110361875, 2025-12-04T09:18:37.5883509Z "digest": "sha256:375c4427e9141269458333b1463fdb219e736fd6231ec1c56c625c48437ace77" 2025-12-04T09:18:37.5883932Z }, 2025-12-04T09:18:37.5884120Z { 2025-12-04T09:18:37.5884441Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5884839Z "size": 4961, 2025-12-04T09:18:37.5885238Z "digest": "sha256:a86faaa7dbdd70e678e5ea20072637ee42618921ca8f80ca089f789325d4b0c2" 2025-12-04T09:18:37.5885693Z }, 2025-12-04T09:18:37.5885881Z { 2025-12-04T09:18:37.5886360Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5886766Z "size": 1755, 2025-12-04T09:18:37.5887158Z "digest": "sha256:fb7848686804957915d98f8655ef6da0fe4c521b50a82aefdebf475983505a15" 2025-12-04T09:18:37.5887680Z }, 2025-12-04T09:18:37.5887866Z { 2025-12-04T09:18:37.5888186Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5888578Z "size": 724, 2025-12-04T09:18:37.5888961Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T09:18:37.5889391Z }, 2025-12-04T09:18:37.5889576Z { 2025-12-04T09:18:37.5889895Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5890292Z "size": 543, 2025-12-04T09:18:37.5890678Z "digest": "sha256:79dc80f426b29d4ae9157b967050b03e66aa0c4b1295b944a1dd70106be87066" 2025-12-04T09:18:37.5891116Z }, 2025-12-04T09:18:37.5891308Z { 2025-12-04T09:18:37.5891635Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5892030Z "size": 3185190117, 2025-12-04T09:18:37.5892452Z "digest": "sha256:a13fcc1b90bb9c251ebe7ef2a03c4cb3afa1c8bdafe84f5f85136773059a3735" 2025-12-04T09:18:37.5892911Z }, 2025-12-04T09:18:37.5893089Z { 2025-12-04T09:18:37.5893409Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5893804Z "size": 32, 2025-12-04T09:18:37.5894196Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.5894638Z }, 2025-12-04T09:18:37.5894825Z { 2025-12-04T09:18:37.5895139Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5895636Z "size": 396, 2025-12-04T09:18:37.5896033Z "digest": "sha256:549db4d6c618ecd9534658a233e3c90508f82d8735f965c2786b2eaa078869e5" 2025-12-04T09:18:37.5896476Z }, 2025-12-04T09:18:37.5896656Z { 2025-12-04T09:18:37.5896987Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5897396Z "size": 236860, 2025-12-04T09:18:37.5897791Z "digest": "sha256:5c63528cb580001e65104f4cb0809bf0673a00f989a7db42fd6d86aa1ec27cee" 2025-12-04T09:18:37.5898236Z }, 2025-12-04T09:18:37.5898430Z { 2025-12-04T09:18:37.5898750Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5899162Z "size": 231, 2025-12-04T09:18:37.5899570Z "digest": "sha256:75bd83b989a44e4d4119a3f972891025eb0e9ce95cfbe4a0ca5cdbe7130028d6" 2025-12-04T09:18:37.5900021Z }, 2025-12-04T09:18:37.5900208Z { 2025-12-04T09:18:37.5900537Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5900941Z "size": 3043497, 2025-12-04T09:18:37.5901330Z "digest": "sha256:de6e78970f517178cb91f36cd02bd9ca7b72a08fb82a0f9007516026f258c035" 2025-12-04T09:18:37.5901771Z }, 2025-12-04T09:18:37.5901960Z { 2025-12-04T09:18:37.5902281Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5902789Z "size": 1472, 2025-12-04T09:18:37.5903206Z "digest": "sha256:e13ed7c7e4736e81dc21af755b3363eb26e4d3b2f1ca988dfe65effa47d8fa42" 2025-12-04T09:18:37.5903649Z }, 2025-12-04T09:18:37.5903834Z { 2025-12-04T09:18:37.5904161Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5904552Z "size": 481, 2025-12-04T09:18:37.5904950Z "digest": "sha256:6e2949bcb74152577a0f20c38bcb6dd80f5e68427e3e531a80e08c9ecc73a979" 2025-12-04T09:18:37.5905393Z }, 2025-12-04T09:18:37.5905582Z { 2025-12-04T09:18:37.5905899Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5906303Z "size": 202, 2025-12-04T09:18:37.5906709Z "digest": "sha256:14d69d9aaec70287efd2fd35c4f93e43a29a4098458cc9fca1c93f02ad7356cb" 2025-12-04T09:18:37.5907151Z }, 2025-12-04T09:18:37.5907341Z { 2025-12-04T09:18:37.5907664Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5908068Z "size": 607, 2025-12-04T09:18:37.5908572Z "digest": "sha256:5c02769dd8e5bba2f7f5fd84bde9595fcb3bdbffcae497503fa846f9b5e78bf5" 2025-12-04T09:18:37.5909035Z }, 2025-12-04T09:18:37.5909215Z { 2025-12-04T09:18:37.5909543Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5909952Z "size": 7889619584, 2025-12-04T09:18:37.5910364Z "digest": "sha256:35041ce524ac4afec40ecd73b1393c830614f1f79d43a6439767a6c7d5b7027b" 2025-12-04T09:18:37.5910811Z }, 2025-12-04T09:18:37.5911007Z { 2025-12-04T09:18:37.5911334Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5911731Z "size": 830, 2025-12-04T09:18:37.5912131Z "digest": "sha256:2fa92dc5885e080e049ceb4139288b6c0e39fab34256945708b08ea55a1f7a0b" 2025-12-04T09:18:37.5912579Z }, 2025-12-04T09:18:37.5912768Z { 2025-12-04T09:18:37.5913090Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5913498Z "size": 33451739, 2025-12-04T09:18:37.5913910Z "digest": "sha256:2b85eafbd92a0e70a0a70154ad8bf4584095e576d95873368f30373f5966714a" 2025-12-04T09:18:37.5914355Z }, 2025-12-04T09:18:37.5914555Z { 2025-12-04T09:18:37.5914877Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5915279Z "size": 104, 2025-12-04T09:18:37.5915687Z "digest": "sha256:ff755a4ddad7880f23c6b767d432d6f1eafdb62b3ea18f8a98e22c441c099fcb" 2025-12-04T09:18:37.5916143Z }, 2025-12-04T09:18:37.5916330Z { 2025-12-04T09:18:37.5916663Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5917071Z "size": 1496, 2025-12-04T09:18:37.5917452Z "digest": "sha256:09eb41bdf42d8605b57b2363348154140904dec914b34a67298b82122bfce2b3" 2025-12-04T09:18:37.5917893Z }, 2025-12-04T09:18:37.5918089Z { 2025-12-04T09:18:37.5918404Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5918809Z "size": 458787828, 2025-12-04T09:18:37.5919216Z "digest": "sha256:11ede4d59e935e62f41b33220fe871794ab5e57ce724173b713368977683bcf6" 2025-12-04T09:18:37.5919658Z }, 2025-12-04T09:18:37.5919844Z { 2025-12-04T09:18:37.5920170Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5920564Z "size": 164, 2025-12-04T09:18:37.5920955Z "digest": "sha256:1283cd8f801a142172f3ab76fd472df8583223d9437de3e4d18d8cf98ea3fa98" 2025-12-04T09:18:37.5921465Z + exit 0 2025-12-04T09:18:37.5921662Z }, 2025-12-04T09:18:37.5921848Z { 2025-12-04T09:18:37.5922167Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5922568Z "size": 346, 2025-12-04T09:18:37.5922953Z "digest": "sha256:024fa855425fa524ad4500660cf61d53be62b99556d31b8b280d14caba434a35" 2025-12-04T09:18:37.5923391Z }, 2025-12-04T09:18:37.5923585Z { 2025-12-04T09:18:37.5923906Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5924308Z "size": 32, 2025-12-04T09:18:37.5924714Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.5925258Z }, 2025-12-04T09:18:37.5925446Z { 2025-12-04T09:18:37.5925769Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5926171Z "size": 106, 2025-12-04T09:18:37.5926567Z "digest": "sha256:303e6747a62efecf5efa1f97d0e66b40a3b39da8d79a51f75b89f4c92ae7ec52" 2025-12-04T09:18:37.5927022Z }, 2025-12-04T09:18:37.5927314Z { 2025-12-04T09:18:37.5927633Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5928052Z "size": 424, 2025-12-04T09:18:37.5928459Z "digest": "sha256:3017cdf4838bcc9a33daebc07487f8ae1f6bd6e7ce8322c14f5480e8db9ef90e" 2025-12-04T09:18:37.5928904Z }, 2025-12-04T09:18:37.5929096Z { 2025-12-04T09:18:37.5929420Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5929816Z "size": 19309374, 2025-12-04T09:18:37.5930230Z "digest": "sha256:6b6cd1c358e886dc6ed7fd46ac4bcc1a0a73b7b1301739ea1953478ee5d83f50" 2025-12-04T09:18:37.5930696Z }, 2025-12-04T09:18:37.5930987Z { 2025-12-04T09:18:37.5931316Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5931727Z "size": 108, 2025-12-04T09:18:37.5932122Z "digest": "sha256:b2dd045011241d1cf8889e2a7369d9fe4844dfe15529b520ccd6a59bd3c1532e" 2025-12-04T09:18:37.5932564Z }, 2025-12-04T09:18:37.5932763Z { 2025-12-04T09:18:37.5933086Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5933488Z "size": 827, 2025-12-04T09:18:37.5933887Z "digest": "sha256:55adc51fe5897031d4cf2f2b8fd162213f6e46a52848630c616606271b97952e" 2025-12-04T09:18:37.5934332Z }, 2025-12-04T09:18:37.5946830Z { 2025-12-04T09:18:37.5947168Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5947559Z "size": 724, 2025-12-04T09:18:37.5947932Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T09:18:37.5948357Z }, 2025-12-04T09:18:37.5948528Z { 2025-12-04T09:18:37.5948836Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5949218Z "size": 149, 2025-12-04T09:18:37.5949590Z "digest": "sha256:a43ca0e4b837964b12b7469194cfe939c26de027298040028975324dce25938a" 2025-12-04T09:18:37.5950004Z }, 2025-12-04T09:18:37.5950173Z { 2025-12-04T09:18:37.5950480Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5950855Z "size": 138, 2025-12-04T09:18:37.5951228Z "digest": "sha256:b7212f17fd1404837fcfdd086dd0e2667931e4db377d45d8d89a44390c84e11d" 2025-12-04T09:18:37.5951651Z }, 2025-12-04T09:18:37.5951819Z { 2025-12-04T09:18:37.5952118Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5952503Z "size": 141, 2025-12-04T09:18:37.5952881Z "digest": "sha256:083e42cac090e6486c35f392b64ee54448f5e4aa947003aeb3e1f92c8ea5c099" 2025-12-04T09:18:37.5953301Z }, 2025-12-04T09:18:37.5953478Z { 2025-12-04T09:18:37.5953798Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5954179Z "size": 32, 2025-12-04T09:18:37.5954568Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.5955005Z }, 2025-12-04T09:18:37.5955180Z { 2025-12-04T09:18:37.5955497Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5955886Z "size": 223, 2025-12-04T09:18:37.5956269Z "digest": "sha256:0a00b784a4aac341795729b254f7edd09e811b7f51d0c58e0e6bfeeee6940503" 2025-12-04T09:18:37.5956705Z }, 2025-12-04T09:18:37.5956887Z { 2025-12-04T09:18:37.5957197Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5957579Z "size": 255, 2025-12-04T09:18:37.5957967Z "digest": "sha256:c6173c779f7ba143a21214ea5f032b141863a37ceb4c0ac01d3248c216ce5241" 2025-12-04T09:18:37.5958396Z }, 2025-12-04T09:18:37.5958573Z { 2025-12-04T09:18:37.5958889Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5959460Z "size": 145520672, 2025-12-04T09:18:37.5959855Z "digest": "sha256:ed3d1e3387b924585c332bf1bc252fa159cd0d25256a874043ff0141b1ab5ff7" 2025-12-04T09:18:37.5960290Z }, 2025-12-04T09:18:37.5960472Z { 2025-12-04T09:18:37.5960780Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5961171Z "size": 106, 2025-12-04T09:18:37.5961544Z "digest": "sha256:b29343478586aeee19d2a622661716f6f1591280c890f49b727a8da13a610784" 2025-12-04T09:18:37.5961970Z }, 2025-12-04T09:18:37.5962150Z { 2025-12-04T09:18:37.5962470Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5962855Z "size": 312293530, 2025-12-04T09:18:37.5963238Z "digest": "sha256:c6f0520487fb506bc4601fd84d5f28d8a76b203e004731e4b2067c2ab1a14e0b" 2025-12-04T09:18:37.5963671Z }, 2025-12-04T09:18:37.5963852Z { 2025-12-04T09:18:37.5964160Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5964563Z "size": 3058011133, 2025-12-04T09:18:37.5965126Z "digest": "sha256:148171691cd4c4d20310d490d4b4dd903490d04ea07fb8f7e668a28768683e9a" 2025-12-04T09:18:37.5965546Z }, 2025-12-04T09:18:37.5965730Z { 2025-12-04T09:18:37.5966053Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5966434Z "size": 129, 2025-12-04T09:18:37.5966827Z "digest": "sha256:2c666d30ed77fff9ff1167d41cd645dad98280fcbe941f5bc3828c7ae66b1287" 2025-12-04T09:18:37.5967329Z }, 2025-12-04T09:18:37.5967507Z { 2025-12-04T09:18:37.5967811Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5968198Z "size": 880, 2025-12-04T09:18:37.5968610Z "digest": "sha256:5d8d3a0a98e012c5068e0f3bae5a03e3148ecf2d063634eee4c9241a1e3fdfb5" 2025-12-04T09:18:37.5969068Z }, 2025-12-04T09:18:37.5969245Z { 2025-12-04T09:18:37.5969558Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5969943Z "size": 724, 2025-12-04T09:18:37.5970321Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T09:18:37.5970746Z }, 2025-12-04T09:18:37.5970921Z { 2025-12-04T09:18:37.5971234Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5971631Z "size": 139, 2025-12-04T09:18:37.5972006Z "digest": "sha256:b06bafce9e817295d8127207747c80aa18e04392ff0875844fc30a1e794a8a0c" 2025-12-04T09:18:37.5972443Z }, 2025-12-04T09:18:37.5972624Z { 2025-12-04T09:18:37.5972941Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5973320Z "size": 32, 2025-12-04T09:18:37.5973709Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.5974147Z }, 2025-12-04T09:18:37.5974316Z { 2025-12-04T09:18:37.5974626Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5975013Z "size": 159, 2025-12-04T09:18:37.5975399Z "digest": "sha256:15e0d7e4590d3d8f598d05aec3a92f891bf8b4605bcc38cc2de852b6014ef8f3" 2025-12-04T09:18:37.5975829Z }, 2025-12-04T09:18:37.5976007Z { 2025-12-04T09:18:37.5976310Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5976694Z "size": 1011, 2025-12-04T09:18:37.5977084Z "digest": "sha256:a514bd1add3164d8d7ca99aa19294c4ed8b97b074635d98714c4f598a959f4cd" 2025-12-04T09:18:37.5977520Z }, 2025-12-04T09:18:37.5977698Z { 2025-12-04T09:18:37.5978022Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5978429Z "size": 724, 2025-12-04T09:18:37.5978828Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T09:18:37.5979262Z }, 2025-12-04T09:18:37.5979448Z { 2025-12-04T09:18:37.5979759Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5980156Z "size": 134, 2025-12-04T09:18:37.5980545Z "digest": "sha256:57b84ee6000204f27a1d9bca199b19be4c86ecd324540dbdf239c56a6c3b34ea" 2025-12-04T09:18:37.5981082Z }, 2025-12-04T09:18:37.5981266Z { 2025-12-04T09:18:37.5981584Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5981967Z "size": 32, 2025-12-04T09:18:37.5982355Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.5982794Z }, 2025-12-04T09:18:37.5982975Z { 2025-12-04T09:18:37.5983284Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5983668Z "size": 157, 2025-12-04T09:18:37.5984070Z "digest": "sha256:b8babeff6d817a5961dddc15c6bdfdbd05da187fae75d5804015f99fd7c066d8" 2025-12-04T09:18:37.5984509Z }, 2025-12-04T09:18:37.5984688Z { 2025-12-04T09:18:37.5985002Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5985383Z "size": 602, 2025-12-04T09:18:37.5985768Z "digest": "sha256:83779ddf6a85ab387f64a45f274cba245b69e4fd1931ff0b5d7d3efd4b7a43bc" 2025-12-04T09:18:37.5986212Z }, 2025-12-04T09:18:37.5986383Z { 2025-12-04T09:18:37.5986795Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5987190Z "size": 724, 2025-12-04T09:18:37.5987555Z "digest": "sha256:3541df015cdb7e8925273399d28e56c31b3c9196f00439ac2925537b173b1f84" 2025-12-04T09:18:37.5987982Z }, 2025-12-04T09:18:37.5988164Z { 2025-12-04T09:18:37.5988486Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5988864Z "size": 155, 2025-12-04T09:18:37.5989252Z "digest": "sha256:8b7620c0d736cc79381207ce5afe2af90f0cd7f0cd394577d2c9520d7f74762f" 2025-12-04T09:18:37.5989683Z }, 2025-12-04T09:18:37.5989856Z { 2025-12-04T09:18:37.5990169Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5990553Z "size": 32, 2025-12-04T09:18:37.5990934Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.5991364Z }, 2025-12-04T09:18:37.5991552Z { 2025-12-04T09:18:37.5991860Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5992246Z "size": 188, 2025-12-04T09:18:37.5992645Z "digest": "sha256:3bcfa090e4efd3677425f76baea9f1e0c50a75d8c6b5713ec05310f1dff24539" 2025-12-04T09:18:37.5993091Z }, 2025-12-04T09:18:37.5993265Z { 2025-12-04T09:18:37.5993585Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5993974Z "size": 1370, 2025-12-04T09:18:37.5994357Z "digest": "sha256:eb0504ec4d9218a79896b604f73dc0ea5a0f96266ad9c2cdbbbe5f0f18222694" 2025-12-04T09:18:37.5994800Z }, 2025-12-04T09:18:37.5994982Z { 2025-12-04T09:18:37.5995291Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5995685Z "size": 32, 2025-12-04T09:18:37.5996080Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.5996507Z }, 2025-12-04T09:18:37.5996679Z { 2025-12-04T09:18:37.5996994Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5997385Z "size": 136, 2025-12-04T09:18:37.5997767Z "digest": "sha256:15d0fec09d7b196a1462d51516ee90fc3443ba178d3e56d59cacf32146b4321d" 2025-12-04T09:18:37.5998202Z }, 2025-12-04T09:18:37.5998379Z { 2025-12-04T09:18:37.5998687Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.5999073Z "size": 528, 2025-12-04T09:18:37.5999468Z "digest": "sha256:cca81fcc62a949959ca4dd3c9056fb293d548ef8607127eeeef6cfd3a8897ca8" 2025-12-04T09:18:37.5999901Z }, 2025-12-04T09:18:37.6000081Z { 2025-12-04T09:18:37.6000386Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6000761Z "size": 32, 2025-12-04T09:18:37.6001148Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.6001588Z }, 2025-12-04T09:18:37.6001758Z { 2025-12-04T09:18:37.6002070Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6002553Z "size": 104, 2025-12-04T09:18:37.6002949Z "digest": "sha256:b0b8f9b5c6ab98db9cd830dc584e1b6aec9add139e4cc48d8c243d36691e25b4" 2025-12-04T09:18:37.6003397Z }, 2025-12-04T09:18:37.6003581Z { 2025-12-04T09:18:37.6003890Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6004272Z "size": 435, 2025-12-04T09:18:37.6004639Z "digest": "sha256:0606ca4d47a8a70e91e92b03ca51a85e731641b09342136a54ef2f2a6d9dfb44" 2025-12-04T09:18:37.6005060Z }, 2025-12-04T09:18:37.6005230Z { 2025-12-04T09:18:37.6005542Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6005930Z "size": 32, 2025-12-04T09:18:37.6006311Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.6006746Z }, 2025-12-04T09:18:37.6006934Z { 2025-12-04T09:18:37.6007289Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6007683Z "size": 109, 2025-12-04T09:18:37.6008166Z "digest": "sha256:2f80a4e1b3b95ed67bb781ea787e8a63e46de79117d9d8e65c257072b38afa2d" 2025-12-04T09:18:37.6008601Z }, 2025-12-04T09:18:37.6008772Z { 2025-12-04T09:18:37.6009086Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6009472Z "size": 1896, 2025-12-04T09:18:37.6009848Z "digest": "sha256:35c916fb1bd057e517dcab78c3a2a018e68096d8993892ad84f47562d37ae352" 2025-12-04T09:18:37.6010279Z }, 2025-12-04T09:18:37.6010461Z { 2025-12-04T09:18:37.6010771Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6011165Z "size": 197526165, 2025-12-04T09:18:37.6011562Z "digest": "sha256:195537b7dafc96192f768323b1a8cc2a914d41959849b73198579576b0872a44" 2025-12-04T09:18:37.6011979Z }, 2025-12-04T09:18:37.6012166Z { 2025-12-04T09:18:37.6012482Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6012871Z "size": 106, 2025-12-04T09:18:37.6013262Z "digest": "sha256:dc454fd3967e5735b2498b7f1d958a2c626987d5e4ce225ca98da3cd945b59f3" 2025-12-04T09:18:37.6013700Z }, 2025-12-04T09:18:37.6013881Z { 2025-12-04T09:18:37.6014182Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6014565Z "size": 165, 2025-12-04T09:18:37.6014947Z "digest": "sha256:701b34f115fa897181c046dc37288e87cbc3ad74c36a9e2224b5bfe7c5703afb" 2025-12-04T09:18:37.6015372Z }, 2025-12-04T09:18:37.6015555Z { 2025-12-04T09:18:37.6015865Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6016246Z "size": 7944, 2025-12-04T09:18:37.6016639Z "digest": "sha256:39cefc00ffedebc9098261c798408b87a20c95a88fccb110594077f48dadf760" 2025-12-04T09:18:37.6017076Z }, 2025-12-04T09:18:37.6017252Z { 2025-12-04T09:18:37.6017566Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6017947Z "size": 8071, 2025-12-04T09:18:37.6018324Z "digest": "sha256:6ae51eb61a325b2c2995a5088c81aa20821b75be65b5aa722c7c40556b5d03ea" 2025-12-04T09:18:37.6018761Z }, 2025-12-04T09:18:37.6018946Z { 2025-12-04T09:18:37.6019256Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6019636Z "size": 304, 2025-12-04T09:18:37.6020024Z "digest": "sha256:1fd5341e66dfc0c1ae23af014641a92a6fd02640c528fe6d4dc55921ed659a26" 2025-12-04T09:18:37.6020463Z }, 2025-12-04T09:18:37.6020637Z { 2025-12-04T09:18:37.6020948Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6021336Z "size": 13364291, 2025-12-04T09:18:37.6021732Z "digest": "sha256:72a7c87e35e40ab796f90aee1b51add7902f0cdc44406d2505b6c6a1f55a8da6" 2025-12-04T09:18:37.6022169Z }, 2025-12-04T09:18:37.6022347Z { 2025-12-04T09:18:37.6022654Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6023044Z "size": 108, 2025-12-04T09:18:37.6023437Z "digest": "sha256:ec36862ac98ebaac52ee1a8b1d162d45bd0e3bf59ae7e19c8f80ad3960b4c600" 2025-12-04T09:18:37.6023979Z }, 2025-12-04T09:18:37.6024159Z { 2025-12-04T09:18:37.6024466Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6024852Z "size": 54145699, 2025-12-04T09:18:37.6025243Z "digest": "sha256:05ddbf246e8add0e293474dbf88bb028d5a295a25ac59e8648a18db644377773" 2025-12-04T09:18:37.6025676Z }, 2025-12-04T09:18:37.6025854Z { 2025-12-04T09:18:37.6026164Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:18:37.6026549Z "size": 32, 2025-12-04T09:18:37.6026933Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:18:37.6027372Z } 2025-12-04T09:18:37.6027569Z ] 2025-12-04T09:18:37.6027774Z } 2025-12-04T09:18:37.6055523Z ##[group]Run set -eux 2025-12-04T09:18:37.6055791Z set -eux 2025-12-04T09:18:37.6056184Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-12-04T09:18:37.6057404Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-12-04T09:18:37.6067009Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:37.6067365Z env: 2025-12-04T09:18:37.6067572Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:37.6067822Z ##[endgroup] 2025-12-04T09:18:37.6101210Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-12-04T09:18:37.6102462Z + jq --raw-output .SecretString 2025-12-04T09:18:37.6103969Z + jq -r .docker_hub_readonly_token 2025-12-04T09:18:37.6105304Z + docker login --username pytorchbot --password-stdin 2025-12-04T09:18:38.1874193Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:18:38.1874862Z Configure a credential helper to remove this warning. See 2025-12-04T09:18:38.1875470Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:18:38.1875945Z 2025-12-04T09:18:38.1876300Z Login Succeeded 2025-12-04T09:18:38.1994492Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T09:18:38.1994870Z tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T09:18:38.1995261Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-12-04T09:18:38.2004890Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:38.2005241Z env: 2025-12-04T09:18:38.2005458Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:38.2006259Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:38.2007062Z ##[endgroup] 2025-12-04T09:18:38.2040021Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:38.2087251Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-12-04T09:18:38.2087679Z with: 2025-12-04T09:18:38.2088413Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:38.2089310Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:38.2089672Z env: 2025-12-04T09:18:38.2089874Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:38.2090127Z ##[endgroup] 2025-12-04T09:18:38.2105179Z ##[group]Run set -x 2025-12-04T09:18:38.2105443Z set -x 2025-12-04T09:18:38.2105664Z set +e 2025-12-04T09:18:38.2105873Z  2025-12-04T09:18:38.2106075Z login() { 2025-12-04T09:18:38.2106540Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T09:18:38.2107034Z } 2025-12-04T09:18:38.2107242Z  2025-12-04T09:18:38.2107476Z retry () { 2025-12-04T09:18:38.2107792Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T09:18:38.2108265Z } 2025-12-04T09:18:38.2108467Z  2025-12-04T09:18:38.2108690Z retry login "${DOCKER_REGISTRY}" 2025-12-04T09:18:38.2108979Z  2025-12-04T09:18:38.2109446Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-12-04T09:18:38.2110078Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-12-04T09:18:38.2110437Z  2025-12-04T09:18:38.2110635Z set -e 2025-12-04T09:18:38.2110964Z # ignore output since only exit code is used for conditional 2025-12-04T09:18:38.2111436Z # only pull docker image if it's not available locally 2025-12-04T09:18:38.2111958Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-12-04T09:18:38.2112441Z  retry docker pull "${DOCKER_IMAGE}" 2025-12-04T09:18:38.2112744Z fi 2025-12-04T09:18:38.2121702Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:18:38.2122071Z env: 2025-12-04T09:18:38.2122274Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:18:38.2123059Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:38.2123955Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:38.2124326Z ##[endgroup] 2025-12-04T09:18:38.2155005Z + set +e 2025-12-04T09:18:38.2155434Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:38.2155850Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:38.2159970Z + aws ecr get-login-password --region us-east-1 2025-12-04T09:18:38.2161209Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:18:38.7460372Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:18:38.7461225Z Configure a credential helper to remove this warning. See 2025-12-04T09:18:38.7461962Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:18:38.7462474Z 2025-12-04T09:18:38.7462646Z Login Succeeded 2025-12-04T09:18:38.7495688Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:38.7496606Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-12-04T09:18:38.9906142Z + IMAGE_SIZE=15091.581844329834 2025-12-04T09:18:38.9906545Z Compressed size of image in MB: 15091.581844329834 2025-12-04T09:18:38.9907095Z + echo 'Compressed size of image in MB: 15091.581844329834' 2025-12-04T09:18:38.9907541Z + set -e 2025-12-04T09:18:38.9908702Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:39.0046187Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:39.0047590Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:18:39.2415832Z pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a: Pulling from pytorch/ci-image 2025-12-04T09:18:39.2417759Z 63e5bc7682b8: Pulling fs layer 2025-12-04T09:18:39.2418206Z 0678d56345c9: Pulling fs layer 2025-12-04T09:18:39.2418583Z 45f5c9ddfce7: Pulling fs layer 2025-12-04T09:18:39.2418987Z 086b1df51ac1: Pulling fs layer 2025-12-04T09:18:39.2419362Z fe8a7b64bf98: Pulling fs layer 2025-12-04T09:18:39.2419730Z 7680723e9a57: Pulling fs layer 2025-12-04T09:18:39.2420138Z 9c5027aeeb4e: Pulling fs layer 2025-12-04T09:18:39.2420528Z 9a5652110360: Pulling fs layer 2025-12-04T09:18:39.2420899Z 375c4427e914: Pulling fs layer 2025-12-04T09:18:39.2421698Z a86faaa7dbdd: Pulling fs layer 2025-12-04T09:18:39.2422083Z fb7848686804: Pulling fs layer 2025-12-04T09:18:39.2422479Z 3541df015cdb: Pulling fs layer 2025-12-04T09:18:39.2422860Z 79dc80f426b2: Pulling fs layer 2025-12-04T09:18:39.2423244Z a13fcc1b90bb: Pulling fs layer 2025-12-04T09:18:39.2423625Z 4f4fb700ef54: Pulling fs layer 2025-12-04T09:18:39.2424039Z 549db4d6c618: Pulling fs layer 2025-12-04T09:18:39.2424395Z 7680723e9a57: Waiting 2025-12-04T09:18:39.2424719Z 5c63528cb580: Pulling fs layer 2025-12-04T09:18:39.2425088Z 75bd83b989a4: Pulling fs layer 2025-12-04T09:18:39.2425442Z 086b1df51ac1: Waiting 2025-12-04T09:18:39.2425771Z de6e78970f51: Pulling fs layer 2025-12-04T09:18:39.2426163Z e13ed7c7e473: Pulling fs layer 2025-12-04T09:18:39.2426521Z 3541df015cdb: Waiting 2025-12-04T09:18:39.2426852Z 6e2949bcb741: Pulling fs layer 2025-12-04T09:18:39.2427233Z fb7848686804: Waiting 2025-12-04T09:18:39.2427568Z 14d69d9aaec7: Pulling fs layer 2025-12-04T09:18:39.2427953Z 9a5652110360: Waiting 2025-12-04T09:18:39.2428315Z 5c02769dd8e5: Pulling fs layer 2025-12-04T09:18:39.2428682Z fe8a7b64bf98: Waiting 2025-12-04T09:18:39.2428958Z 375c4427e914: Waiting 2025-12-04T09:18:39.2429188Z 35041ce524ac: Pulling fs layer 2025-12-04T09:18:39.2429527Z e13ed7c7e473: Waiting 2025-12-04T09:18:39.2429793Z 2fa92dc5885e: Pulling fs layer 2025-12-04T09:18:39.2430059Z 2b85eafbd92a: Pulling fs layer 2025-12-04T09:18:39.2430331Z ff755a4ddad7: Pulling fs layer 2025-12-04T09:18:39.2430594Z de6e78970f51: Waiting 2025-12-04T09:18:39.2430825Z 09eb41bdf42d: Pulling fs layer 2025-12-04T09:18:39.2431092Z 11ede4d59e93: Pulling fs layer 2025-12-04T09:18:39.2431362Z 1283cd8f801a: Pulling fs layer 2025-12-04T09:18:39.2431622Z 024fa855425f: Pulling fs layer 2025-12-04T09:18:39.2431874Z 79dc80f426b2: Waiting 2025-12-04T09:18:39.2432101Z 2fa92dc5885e: Waiting 2025-12-04T09:18:39.2432330Z 75bd83b989a4: Waiting 2025-12-04T09:18:39.2432552Z 5c63528cb580: Waiting 2025-12-04T09:18:39.2432779Z a86faaa7dbdd: Waiting 2025-12-04T09:18:39.2433015Z a13fcc1b90bb: Waiting 2025-12-04T09:18:39.2433239Z 4f4fb700ef54: Waiting 2025-12-04T09:18:39.2433461Z 35041ce524ac: Waiting 2025-12-04T09:18:39.2433684Z 549db4d6c618: Waiting 2025-12-04T09:18:39.2433906Z 09eb41bdf42d: Waiting 2025-12-04T09:18:39.2434129Z 1283cd8f801a: Waiting 2025-12-04T09:18:39.2434364Z 303e6747a62e: Pulling fs layer 2025-12-04T09:18:39.2434623Z 3017cdf4838b: Pulling fs layer 2025-12-04T09:18:39.2434881Z 2b85eafbd92a: Waiting 2025-12-04T09:18:39.2435105Z 024fa855425f: Waiting 2025-12-04T09:18:39.2435623Z 6b6cd1c358e8: Pulling fs layer 2025-12-04T09:18:39.2435878Z 303e6747a62e: Waiting 2025-12-04T09:18:39.2436113Z b2dd04501124: Pulling fs layer 2025-12-04T09:18:39.2436380Z 55adc51fe589: Pulling fs layer 2025-12-04T09:18:39.2436636Z 3017cdf4838b: Waiting 2025-12-04T09:18:39.2436902Z 11ede4d59e93: Waiting 2025-12-04T09:18:39.2437202Z b2dd04501124: Waiting 2025-12-04T09:18:39.2437790Z a43ca0e4b837: Pulling fs layer 2025-12-04T09:18:39.2438191Z 55adc51fe589: Waiting 2025-12-04T09:18:39.2438566Z b7212f17fd14: Pulling fs layer 2025-12-04T09:18:39.2438968Z 083e42cac090: Pulling fs layer 2025-12-04T09:18:39.2439321Z 5c02769dd8e5: Waiting 2025-12-04T09:18:39.2439651Z 0a00b784a4aa: Pulling fs layer 2025-12-04T09:18:39.2440004Z b7212f17fd14: Waiting 2025-12-04T09:18:39.2440298Z 083e42cac090: Waiting 2025-12-04T09:18:39.2440579Z c6173c779f7b: Pulling fs layer 2025-12-04T09:18:39.2440842Z 0a00b784a4aa: Waiting 2025-12-04T09:18:39.2441158Z ff755a4ddad7: Waiting 2025-12-04T09:18:39.2441424Z 14d69d9aaec7: Waiting 2025-12-04T09:18:39.2441644Z 6e2949bcb741: Waiting 2025-12-04T09:18:39.2441884Z ed3d1e3387b9: Pulling fs layer 2025-12-04T09:18:39.2442266Z b29343478586: Pulling fs layer 2025-12-04T09:18:39.2442603Z c6f0520487fb: Pulling fs layer 2025-12-04T09:18:39.2442856Z ed3d1e3387b9: Waiting 2025-12-04T09:18:39.2443091Z 148171691cd4: Pulling fs layer 2025-12-04T09:18:39.2454447Z c6f0520487fb: Waiting 2025-12-04T09:18:39.2454718Z 2c666d30ed77: Pulling fs layer 2025-12-04T09:18:39.2455208Z a43ca0e4b837: Waiting 2025-12-04T09:18:39.2455467Z 5d8d3a0a98e0: Pulling fs layer 2025-12-04T09:18:39.2455737Z 148171691cd4: Waiting 2025-12-04T09:18:39.2455965Z 2c666d30ed77: Waiting 2025-12-04T09:18:39.2456218Z b06bafce9e81: Pulling fs layer 2025-12-04T09:18:39.2456493Z b06bafce9e81: Waiting 2025-12-04T09:18:39.2456741Z 15e0d7e4590d: Pulling fs layer 2025-12-04T09:18:39.2457007Z c6173c779f7b: Waiting 2025-12-04T09:18:39.2457260Z a514bd1add31: Pulling fs layer 2025-12-04T09:18:39.2457517Z 15e0d7e4590d: Waiting 2025-12-04T09:18:39.2457767Z 57b84ee60002: Pulling fs layer 2025-12-04T09:18:39.2458091Z b8babeff6d81: Pulling fs layer 2025-12-04T09:18:39.2458369Z a514bd1add31: Waiting 2025-12-04T09:18:39.2458620Z 83779ddf6a85: Pulling fs layer 2025-12-04T09:18:39.2458884Z 57b84ee60002: Waiting 2025-12-04T09:18:39.2459112Z 8b7620c0d736: Pulling fs layer 2025-12-04T09:18:39.2459380Z 5d8d3a0a98e0: Waiting 2025-12-04T09:18:39.2459639Z 3bcfa090e4ef: Pulling fs layer 2025-12-04T09:18:39.2459914Z eb0504ec4d92: Pulling fs layer 2025-12-04T09:18:39.2460201Z 15d0fec09d7b: Pulling fs layer 2025-12-04T09:18:39.2460470Z 83779ddf6a85: Waiting 2025-12-04T09:18:39.2460696Z 8b7620c0d736: Waiting 2025-12-04T09:18:39.2460936Z 3bcfa090e4ef: Waiting 2025-12-04T09:18:39.2461188Z cca81fcc62a9: Pulling fs layer 2025-12-04T09:18:39.2461453Z eb0504ec4d92: Waiting 2025-12-04T09:18:39.2461681Z cca81fcc62a9: Waiting 2025-12-04T09:18:39.2461942Z b0b8f9b5c6ab: Pulling fs layer 2025-12-04T09:18:39.2462200Z b29343478586: Waiting 2025-12-04T09:18:39.2462438Z 0606ca4d47a8: Pulling fs layer 2025-12-04T09:18:39.2462718Z 2f80a4e1b3b9: Pulling fs layer 2025-12-04T09:18:39.2462999Z 35c916fb1bd0: Pulling fs layer 2025-12-04T09:18:39.2463252Z 0606ca4d47a8: Waiting 2025-12-04T09:18:39.2463500Z 195537b7dafc: Pulling fs layer 2025-12-04T09:18:39.2463766Z 2f80a4e1b3b9: Waiting 2025-12-04T09:18:39.2463996Z b0b8f9b5c6ab: Waiting 2025-12-04T09:18:39.2464246Z dc454fd3967e: Pulling fs layer 2025-12-04T09:18:39.2464510Z 9c5027aeeb4e: Waiting 2025-12-04T09:18:39.2464753Z 701b34f115fa: Pulling fs layer 2025-12-04T09:18:39.2465078Z 195537b7dafc: Waiting 2025-12-04T09:18:39.2465388Z 39cefc00ffed: Pulling fs layer 2025-12-04T09:18:39.2465745Z dc454fd3967e: Waiting 2025-12-04T09:18:39.2466103Z 6ae51eb61a32: Pulling fs layer 2025-12-04T09:18:39.2466513Z 1fd5341e66df: Pulling fs layer 2025-12-04T09:18:39.2466905Z 39cefc00ffed: Waiting 2025-12-04T09:18:39.2467230Z 72a7c87e35e4: Pulling fs layer 2025-12-04T09:18:39.2467578Z ec36862ac98e: Pulling fs layer 2025-12-04T09:18:39.2467841Z 1fd5341e66df: Waiting 2025-12-04T09:18:39.2468064Z 6ae51eb61a32: Waiting 2025-12-04T09:18:39.2468298Z 35c916fb1bd0: Waiting 2025-12-04T09:18:39.2468545Z 05ddbf246e8a: Pulling fs layer 2025-12-04T09:18:39.2468794Z ec36862ac98e: Waiting 2025-12-04T09:18:39.2469026Z 05ddbf246e8a: Waiting 2025-12-04T09:18:39.2469246Z 72a7c87e35e4: Waiting 2025-12-04T09:18:39.2469548Z 15d0fec09d7b: Waiting 2025-12-04T09:18:39.3373866Z 0678d56345c9: Download complete 2025-12-04T09:18:39.4305037Z 086b1df51ac1: Download complete 2025-12-04T09:18:39.5224233Z fe8a7b64bf98: Verifying Checksum 2025-12-04T09:18:39.5224554Z fe8a7b64bf98: Download complete 2025-12-04T09:18:39.6068249Z 7680723e9a57: Verifying Checksum 2025-12-04T09:18:39.6068713Z 7680723e9a57: Download complete 2025-12-04T09:18:39.6142423Z 63e5bc7682b8: Verifying Checksum 2025-12-04T09:18:39.6142798Z 63e5bc7682b8: Download complete 2025-12-04T09:18:39.7047152Z 9a5652110360: Download complete 2025-12-04T09:18:39.7056487Z 9c5027aeeb4e: Verifying Checksum 2025-12-04T09:18:39.7057089Z 9c5027aeeb4e: Download complete 2025-12-04T09:18:39.7723491Z a86faaa7dbdd: Download complete 2025-12-04T09:18:39.8361712Z fb7848686804: Download complete 2025-12-04T09:18:39.9140652Z 3541df015cdb: Download complete 2025-12-04T09:18:39.9996769Z 79dc80f426b2: Download complete 2025-12-04T09:18:40.8029380Z 63e5bc7682b8: Pull complete 2025-12-04T09:18:40.8140748Z 0678d56345c9: Pull complete 2025-12-04T09:18:40.8634374Z 375c4427e914: Verifying Checksum 2025-12-04T09:18:40.8634672Z 375c4427e914: Download complete 2025-12-04T09:18:40.8720082Z 4f4fb700ef54: Download complete 2025-12-04T09:18:40.9425993Z 549db4d6c618: Verifying Checksum 2025-12-04T09:18:40.9426290Z 549db4d6c618: Download complete 2025-12-04T09:18:41.0208947Z 5c63528cb580: Verifying Checksum 2025-12-04T09:18:41.0209246Z 5c63528cb580: Download complete 2025-12-04T09:18:41.2146194Z de6e78970f51: Verifying Checksum 2025-12-04T09:18:41.2146631Z de6e78970f51: Download complete 2025-12-04T09:18:41.2731248Z e13ed7c7e473: Download complete 2025-12-04T09:18:41.3533427Z 6e2949bcb741: Verifying Checksum 2025-12-04T09:18:41.3533846Z 6e2949bcb741: Download complete 2025-12-04T09:18:41.4122388Z 14d69d9aaec7: Verifying Checksum 2025-12-04T09:18:41.4122812Z 14d69d9aaec7: Download complete 2025-12-04T09:18:41.4804144Z 5c02769dd8e5: Verifying Checksum 2025-12-04T09:18:41.4804561Z 5c02769dd8e5: Download complete 2025-12-04T09:18:42.4240177Z 45f5c9ddfce7: Verifying Checksum 2025-12-04T09:18:42.4240511Z 45f5c9ddfce7: Download complete 2025-12-04T09:18:42.5113768Z 2fa92dc5885e: Verifying Checksum 2025-12-04T09:18:42.5114233Z 2fa92dc5885e: Download complete 2025-12-04T09:18:42.8977616Z 2b85eafbd92a: Verifying Checksum 2025-12-04T09:18:42.8977963Z 2b85eafbd92a: Download complete 2025-12-04T09:18:42.9779960Z ff755a4ddad7: Verifying Checksum 2025-12-04T09:18:42.9780412Z ff755a4ddad7: Download complete 2025-12-04T09:18:43.0677193Z 09eb41bdf42d: Download complete 2025-12-04T09:18:47.7162867Z 11ede4d59e93: Verifying Checksum 2025-12-04T09:18:47.7163335Z 11ede4d59e93: Download complete 2025-12-04T09:18:47.8148182Z 1283cd8f801a: Download complete 2025-12-04T09:18:47.8842707Z 024fa855425f: Verifying Checksum 2025-12-04T09:18:47.8843042Z 024fa855425f: Download complete 2025-12-04T09:18:47.9853722Z 303e6747a62e: Verifying Checksum 2025-12-04T09:18:47.9854050Z 303e6747a62e: Download complete 2025-12-04T09:18:48.0499172Z 3017cdf4838b: Verifying Checksum 2025-12-04T09:18:48.0499560Z 3017cdf4838b: Download complete 2025-12-04T09:18:48.3002889Z 6b6cd1c358e8: Verifying Checksum 2025-12-04T09:18:48.3003328Z 6b6cd1c358e8: Download complete 2025-12-04T09:18:48.3972601Z b2dd04501124: Verifying Checksum 2025-12-04T09:18:48.3972925Z b2dd04501124: Download complete 2025-12-04T09:18:48.5126919Z 55adc51fe589: Verifying Checksum 2025-12-04T09:18:48.5127435Z 55adc51fe589: Download complete 2025-12-04T09:18:48.6015267Z a43ca0e4b837: Verifying Checksum 2025-12-04T09:18:48.6015712Z a43ca0e4b837: Download complete 2025-12-04T09:18:48.6775421Z b7212f17fd14: Verifying Checksum 2025-12-04T09:18:48.7480656Z 083e42cac090: Verifying Checksum 2025-12-04T09:18:48.7481002Z 083e42cac090: Download complete 2025-12-04T09:18:48.8269196Z 0a00b784a4aa: Download complete 2025-12-04T09:18:48.8916603Z c6173c779f7b: Download complete 2025-12-04T09:18:50.3943212Z ed3d1e3387b9: Verifying Checksum 2025-12-04T09:18:50.3943645Z ed3d1e3387b9: Download complete 2025-12-04T09:18:50.4827678Z b29343478586: Download complete 2025-12-04T09:18:52.4367250Z 45f5c9ddfce7: Pull complete 2025-12-04T09:18:52.6480069Z 086b1df51ac1: Pull complete 2025-12-04T09:18:52.8057391Z fe8a7b64bf98: Pull complete 2025-12-04T09:18:52.9820296Z 7680723e9a57: Pull complete 2025-12-04T09:18:53.1246052Z 9c5027aeeb4e: Pull complete 2025-12-04T09:18:53.3009761Z 9a5652110360: Pull complete 2025-12-04T09:18:53.7074323Z c6f0520487fb: Verifying Checksum 2025-12-04T09:18:53.7074884Z c6f0520487fb: Download complete 2025-12-04T09:18:55.9544690Z 375c4427e914: Pull complete 2025-12-04T09:18:56.1804792Z a86faaa7dbdd: Pull complete 2025-12-04T09:18:56.3869677Z fb7848686804: Pull complete 2025-12-04T09:18:56.5884675Z 3541df015cdb: Pull complete 2025-12-04T09:18:56.8004478Z 79dc80f426b2: Pull complete 2025-12-04T09:19:11.9017383Z a13fcc1b90bb: Verifying Checksum 2025-12-04T09:19:11.9017793Z a13fcc1b90bb: Download complete 2025-12-04T09:19:11.9734791Z 2c666d30ed77: Verifying Checksum 2025-12-04T09:19:11.9735501Z 2c666d30ed77: Download complete 2025-12-04T09:19:12.0773462Z 5d8d3a0a98e0: Verifying Checksum 2025-12-04T09:19:12.0773870Z 5d8d3a0a98e0: Download complete 2025-12-04T09:19:12.1600614Z b06bafce9e81: Verifying Checksum 2025-12-04T09:19:12.1600956Z b06bafce9e81: Download complete 2025-12-04T09:19:12.2516458Z 15e0d7e4590d: Verifying Checksum 2025-12-04T09:19:12.2516882Z 15e0d7e4590d: Download complete 2025-12-04T09:19:12.3639520Z a514bd1add31: Verifying Checksum 2025-12-04T09:19:12.3640099Z a514bd1add31: Download complete 2025-12-04T09:19:12.4286809Z 57b84ee60002: Verifying Checksum 2025-12-04T09:19:12.4287140Z 57b84ee60002: Download complete 2025-12-04T09:19:12.5185477Z b8babeff6d81: Verifying Checksum 2025-12-04T09:19:12.5185933Z b8babeff6d81: Download complete 2025-12-04T09:19:12.6033321Z 83779ddf6a85: Verifying Checksum 2025-12-04T09:19:12.6033795Z 83779ddf6a85: Download complete 2025-12-04T09:19:12.6770809Z 8b7620c0d736: Verifying Checksum 2025-12-04T09:19:12.6771242Z 8b7620c0d736: Download complete 2025-12-04T09:19:12.7682183Z 3bcfa090e4ef: Verifying Checksum 2025-12-04T09:19:12.7682627Z 3bcfa090e4ef: Download complete 2025-12-04T09:19:12.8433399Z eb0504ec4d92: Verifying Checksum 2025-12-04T09:19:12.8433840Z eb0504ec4d92: Download complete 2025-12-04T09:19:12.9117454Z 15d0fec09d7b: Verifying Checksum 2025-12-04T09:19:12.9117873Z 15d0fec09d7b: Download complete 2025-12-04T09:19:12.9950448Z cca81fcc62a9: Verifying Checksum 2025-12-04T09:19:12.9950877Z cca81fcc62a9: Download complete 2025-12-04T09:19:13.0806453Z b0b8f9b5c6ab: Verifying Checksum 2025-12-04T09:19:13.0806820Z b0b8f9b5c6ab: Download complete 2025-12-04T09:19:13.2950811Z 0606ca4d47a8: Download complete 2025-12-04T09:19:13.4065156Z 2f80a4e1b3b9: Verifying Checksum 2025-12-04T09:19:13.4065509Z 2f80a4e1b3b9: Download complete 2025-12-04T09:19:13.4792763Z 35c916fb1bd0: Download complete 2025-12-04T09:19:15.5208056Z 195537b7dafc: Verifying Checksum 2025-12-04T09:19:15.5208512Z 195537b7dafc: Download complete 2025-12-04T09:19:15.6052985Z dc454fd3967e: Verifying Checksum 2025-12-04T09:19:15.6053495Z dc454fd3967e: Download complete 2025-12-04T09:19:15.6824718Z 701b34f115fa: Download complete 2025-12-04T09:19:15.7857960Z 39cefc00ffed: Verifying Checksum 2025-12-04T09:19:15.7858560Z 39cefc00ffed: Download complete 2025-12-04T09:19:15.8798975Z 6ae51eb61a32: Verifying Checksum 2025-12-04T09:19:15.8799297Z 6ae51eb61a32: Download complete 2025-12-04T09:19:15.9584499Z 1fd5341e66df: Verifying Checksum 2025-12-04T09:19:15.9584817Z 1fd5341e66df: Download complete 2025-12-04T09:19:16.1408031Z 72a7c87e35e4: Verifying Checksum 2025-12-04T09:19:16.1408509Z 72a7c87e35e4: Download complete 2025-12-04T09:19:16.2226233Z ec36862ac98e: Verifying Checksum 2025-12-04T09:19:16.2227620Z ec36862ac98e: Download complete 2025-12-04T09:19:16.8405021Z 05ddbf246e8a: Verifying Checksum 2025-12-04T09:19:16.8405368Z 05ddbf246e8a: Download complete 2025-12-04T09:19:24.3364162Z 148171691cd4: Verifying Checksum 2025-12-04T09:19:24.3364558Z 148171691cd4: Download complete 2025-12-04T09:20:00.4725700Z 35041ce524ac: Verifying Checksum 2025-12-04T09:20:00.4726406Z 35041ce524ac: Download complete 2025-12-04T09:20:35.2131298Z a13fcc1b90bb: Pull complete 2025-12-04T09:20:35.3844862Z 4f4fb700ef54: Pull complete 2025-12-04T09:20:35.5885889Z 549db4d6c618: Pull complete 2025-12-04T09:20:35.8292636Z 5c63528cb580: Pull complete 2025-12-04T09:20:35.9483193Z 75bd83b989a4: Pull complete 2025-12-04T09:20:36.1454245Z de6e78970f51: Pull complete 2025-12-04T09:20:36.3669930Z e13ed7c7e473: Pull complete 2025-12-04T09:20:36.5848121Z 6e2949bcb741: Pull complete 2025-12-04T09:20:36.7338130Z 14d69d9aaec7: Pull complete 2025-12-04T09:20:36.8876696Z 5c02769dd8e5: Pull complete 2025-12-04T09:22:07.8590544Z 35041ce524ac: Pull complete 2025-12-04T09:22:08.0673189Z 2fa92dc5885e: Pull complete 2025-12-04T09:22:08.7795676Z 2b85eafbd92a: Pull complete 2025-12-04T09:22:08.9758764Z ff755a4ddad7: Pull complete 2025-12-04T09:22:09.0004064Z 09eb41bdf42d: Pull complete 2025-12-04T09:22:17.6407933Z 11ede4d59e93: Pull complete 2025-12-04T09:22:17.8638325Z 1283cd8f801a: Pull complete 2025-12-04T09:22:18.0765910Z 024fa855425f: Pull complete 2025-12-04T09:22:18.4858656Z 303e6747a62e: Pull complete 2025-12-04T09:22:18.7090299Z 3017cdf4838b: Pull complete 2025-12-04T09:22:19.1283864Z 6b6cd1c358e8: Pull complete 2025-12-04T09:22:19.3396845Z b2dd04501124: Pull complete 2025-12-04T09:22:19.5474458Z 55adc51fe589: Pull complete 2025-12-04T09:22:19.9875825Z a43ca0e4b837: Pull complete 2025-12-04T09:22:20.2056684Z b7212f17fd14: Pull complete 2025-12-04T09:22:20.4185913Z 083e42cac090: Pull complete 2025-12-04T09:22:20.8609586Z 0a00b784a4aa: Pull complete 2025-12-04T09:22:21.0694747Z c6173c779f7b: Pull complete 2025-12-04T09:22:24.8158792Z ed3d1e3387b9: Pull complete 2025-12-04T09:22:25.0275153Z b29343478586: Pull complete 2025-12-04T09:22:26.4512353Z c6f0520487fb: Pull complete 2025-12-04T09:23:27.6781194Z 148171691cd4: Pull complete 2025-12-04T09:23:27.7625617Z 2c666d30ed77: Pull complete 2025-12-04T09:23:27.9340979Z 5d8d3a0a98e0: Pull complete 2025-12-04T09:23:28.2912095Z b06bafce9e81: Pull complete 2025-12-04T09:23:28.3963385Z 15e0d7e4590d: Pull complete 2025-12-04T09:23:28.4211708Z a514bd1add31: Pull complete 2025-12-04T09:23:28.6220246Z 57b84ee60002: Pull complete 2025-12-04T09:23:28.8485636Z b8babeff6d81: Pull complete 2025-12-04T09:23:28.9471194Z 83779ddf6a85: Pull complete 2025-12-04T09:23:29.3769550Z 8b7620c0d736: Pull complete 2025-12-04T09:23:29.7988840Z 3bcfa090e4ef: Pull complete 2025-12-04T09:23:30.0073403Z eb0504ec4d92: Pull complete 2025-12-04T09:23:30.4373331Z 15d0fec09d7b: Pull complete 2025-12-04T09:23:30.6100994Z cca81fcc62a9: Pull complete 2025-12-04T09:23:30.8989341Z b0b8f9b5c6ab: Pull complete 2025-12-04T09:23:31.0341368Z 0606ca4d47a8: Pull complete 2025-12-04T09:23:31.3678093Z 2f80a4e1b3b9: Pull complete 2025-12-04T09:23:31.4059539Z 35c916fb1bd0: Pull complete 2025-12-04T09:23:38.2337928Z 195537b7dafc: Pull complete 2025-12-04T09:23:38.4698820Z dc454fd3967e: Pull complete 2025-12-04T09:23:38.6819043Z 701b34f115fa: Pull complete 2025-12-04T09:23:38.9073355Z 39cefc00ffed: Pull complete 2025-12-04T09:23:39.0819393Z 6ae51eb61a32: Pull complete 2025-12-04T09:23:39.2140339Z 1fd5341e66df: Pull complete 2025-12-04T09:23:40.9216584Z 72a7c87e35e4: Pull complete 2025-12-04T09:23:41.0236081Z ec36862ac98e: Pull complete 2025-12-04T09:23:42.7515331Z 05ddbf246e8a: Pull complete 2025-12-04T09:23:43.0730363Z Digest: sha256:ba21003510dba4bdeed83df81a56fa468e0ee1b612a9445ae1f402a280804f97 2025-12-04T09:23:43.1109252Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:23:43.1322281Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:23:43.1408206Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:23:43.1409267Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:23:43.1420123Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:23:43.1420480Z env: 2025-12-04T09:23:43.1420698Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:23:43.1420960Z ##[endgroup] 2025-12-04T09:23:43.1621373Z ##[group]Run pytorch/test-infra/.github/actions/setup-nvidia@main 2025-12-04T09:23:43.1621779Z with: 2025-12-04T09:23:43.1622001Z driver-version: 580.82.07 2025-12-04T09:23:43.1622247Z env: 2025-12-04T09:23:43.1622456Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:23:43.1622712Z ##[endgroup] 2025-12-04T09:23:43.1742820Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:23:43.1743670Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:23:43.1754103Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:23:43.1754461Z env: 2025-12-04T09:23:43.1754672Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:23:43.1755113Z ##[endgroup] 2025-12-04T09:23:43.1828445Z ##[group]Run set -euo pipefail 2025-12-04T09:23:43.1828763Z set -euo pipefail 2025-12-04T09:23:43.1829053Z  2025-12-04T09:23:43.1829266Z has_gpu=false 2025-12-04T09:23:43.1829514Z devices="" 2025-12-04T09:23:43.1829735Z  2025-12-04T09:23:43.1830008Z if command -v nvidia-smi >/dev/null 2>&1; then 2025-12-04T09:23:43.1830438Z  if nvidia-smi -L >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:23:43.1830807Z  has_gpu=true 2025-12-04T09:23:43.1831086Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:23:43.1831395Z  fi 2025-12-04T09:23:43.1831603Z fi 2025-12-04T09:23:43.1831795Z  2025-12-04T09:23:43.1832017Z if [ "$has_gpu" = false ]; then 2025-12-04T09:23:43.1832396Z  if ls /dev/nvidia* >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:23:43.1832756Z  has_gpu=true 2025-12-04T09:23:43.1833049Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:23:43.1833357Z  fi 2025-12-04T09:23:43.1833565Z fi 2025-12-04T09:23:43.1833762Z  2025-12-04T09:23:43.1834069Z if [ "$has_gpu" = false ] && command -v lspci >/dev/null 2>&1; then 2025-12-04T09:23:43.1834563Z  if lspci | grep -i 'nvidia' >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:23:43.1834957Z  has_gpu=true 2025-12-04T09:23:43.1835647Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:23:43.1835957Z  fi 2025-12-04T09:23:43.1836160Z fi 2025-12-04T09:23:43.1836361Z  2025-12-04T09:23:43.1836664Z printf 'HAS_NVIDIA=%s\n' "$has_gpu" >> "$GITHUB_OUTPUT" 2025-12-04T09:23:43.1837181Z printf 'DETECTED_DEVICES<> "$GITHUB_OUTPUT" 2025-12-04T09:23:43.1845682Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:23:43.1846041Z env: 2025-12-04T09:23:43.1846251Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:23:43.1846499Z ##[endgroup] 2025-12-04T09:23:44.9253983Z ##[group]Run if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:23:44.9254370Z if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:23:44.9254725Z  echo "HAS_NVIDIA_GPU=true" >> "${GITHUB_ENV}" 2025-12-04T09:23:44.9255226Z  echo "GPU_FLAG=--gpus all -e NVIDIA_DRIVER_CAPABILITIES=all" >> "${GITHUB_ENV}" 2025-12-04T09:23:44.9255667Z else 2025-12-04T09:23:44.9255939Z  echo "HAS_NVIDIA_GPU=false" >> "${GITHUB_ENV}" 2025-12-04T09:23:44.9256268Z fi 2025-12-04T09:23:44.9265745Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:23:44.9266102Z env: 2025-12-04T09:23:44.9266319Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:23:44.9266573Z HAS_NVIDIA: true 2025-12-04T09:23:44.9266798Z ##[endgroup] 2025-12-04T09:23:44.9346095Z ##[group]Run nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482 2025-12-04T09:23:44.9346503Z with: 2025-12-04T09:23:44.9346704Z timeout_minutes: 10 2025-12-04T09:23:44.9346976Z max_attempts: 3 2025-12-04T09:23:44.9371813Z command: # Is it disgusting to have a full shell script here in this github action? Sure # But is it the best way to make it so that this action relies on nothing else? Absolutely set -eou pipefail DISTRIBUTION=$(. /etc/os-release;echo $ID$VERSION_ID) DRIVER_FN="NVIDIA-Linux-x86_64-${DRIVER_VERSION}.run" install_nvidia_docker2_amzn2() { ( set -x # Needed for yum-config-manager sudo yum install -y yum-utils if [[ "${DISTRIBUTION}" == "amzn2023" ]] ; then YUM_REPO_URL="https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo" else # Amazon Linux 2 YUM_REPO_URL="https://nvidia.github.io/nvidia-docker/${DISTRIBUTION}/nvidia-docker.repo" fi sudo yum-config-manager --add-repo "${YUM_REPO_URL}" sudo yum install -y \ nvidia-container-toolkit-1.17.8 \ libnvidia-container-tools-1.17.8 \ libnvidia-container1-1.17.8 \ nvidia-container-toolkit-base-1.17.8 sudo systemctl restart docker ) } install_nvidia_docker2_ubuntu20() { ( set -x # Install nvidia-driver package if not installed status="$(dpkg-query -W --showformat='${db:Status-Status}' nvidia-docker2 2>&1)" if [ ! $? = 0 ] || [ ! "$status" = installed ]; then sudo apt-get install -y nvidia-container-toolkit-1.17.8 sudo systemctl restart docker fi ) } pre_install_nvidia_driver_amzn2() { ( # Purge any nvidia driver installed from RHEL repo sudo yum remove -y nvidia-driver-latest-dkms ) } install_nvidia_driver_common() { ( # Try to gather more information about the runner and its existing NVIDIA driver if any echo "Before installing NVIDIA driver" lspci lsmod modinfo nvidia || true HAS_NVIDIA_DRIVER=0 # Check if NVIDIA driver has already been installed if [ -x "$(command -v nvidia-smi)" ]; then set +e # The driver exists, check its version next. Also check only the first GPU if there are more than one of them # so that the same driver version is not print over multiple lines INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then echo "Failed to get NVIDIA driver version ($INSTALLED_DRIVER_VERSION). Continuing" elif [ "$INSTALLED_DRIVER_VERSION" != "$DRIVER_VERSION" ]; then echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has been installed, but we expect to have $DRIVER_VERSION instead. Continuing" # Turn off persistent mode so that the installation script can unload the kernel module sudo killall nvidia-persistenced || true else HAS_NVIDIA_DRIVER=1 echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has already been installed. Skipping NVIDIA driver installation" fi set -e fi if [ "$HAS_NVIDIA_DRIVER" -eq 0 ]; then # CAUTION: this may need to be updated in future if [ "${DISTRIBUTION}" != ubuntu20.04 ]; then sudo yum groupinstall -y "Development Tools" # ensure our kernel install is the same as our underlying kernel, # groupinstall "Development Tools" has a habit of mismatching kernel headers sudo yum install -y "kernel-devel-uname-r == $(uname -r)" sudo modprobe backlight fi sudo curl -fsL -o /tmp/nvidia_driver "https://s3.amazonaws.com/ossci-linux/nvidia_driver/$DRIVER_FN" set +e sudo /bin/bash /tmp/nvidia_driver -s --no-drm NVIDIA_INSTALLATION_STATUS=$? RESET_GPU=0 if [ "$NVIDIA_INSTALLATION_STATUS" -ne 0 ]; then sudo cat /var/log/nvidia-installer.log # Fail to install NVIDIA driver, try to reset the GPU RESET_GPU=1 elif [ -x "$(command -v nvidia-smi)" ]; then # Check again if nvidia-smi works even if the driver installation completes successfully INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then RESET_GPU=1 fi fi if [ "$RESET_GPU" -eq 1 ]; then NVIDIA_DEVICES=$(lspci -D | grep -i NVIDIA | cut -d' ' -f1) # The GPU can get stuck in a failure state if somehow the test crashs the GPU microcode. When this # happens, we'll try to reset all NVIDIA devices https://github.com/pytorch/pytorch/issues/88388 for PCI_ID in $NVIDIA_DEVICES; do DEVICE_ENABLED=$(cat /sys/bus/pci/devices/$PCI_ID/enable) echo "Reseting $PCI_ID (enabled state: $DEVICE_ENABLED)" # This requires sudo permission of course echo "1" | sudo tee /sys/bus/pci/devices/$PCI_ID/reset sleep 1 done fi sudo rm -fv /tmp/nvidia_driver set -e fi ) } post_install_nvidia_driver_common() { ( sudo modprobe nvidia || true echo "After installing NVIDIA driver" lspci lsmod modinfo nvidia || true ( set +e nvidia-smi # NB: Annoyingly, nvidia-smi command returns successfully with return code 0 even in # the case where the driver has already crashed as it still can get the driver version # and some basic information like the bus ID. However, the rest of the information # would be missing (ERR!), for example: # # +-----------------------------------------------------------------------------+ # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | # |-------------------------------+----------------------+----------------------+ # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | # | | | MIG M. | # |===============================+======================+======================| # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | # | | | ERR! | # +-------------------------------+----------------------+----------------------+ # # +-----------------------------------------------------------------------------+ # | Processes: | # | GPU GI CI PID Type Process name GPU Memory | # | ID ID Usage | # |=============================================================================| # +-----------------------------------------------------------------------------+ # # This should be reported as a failure instead as it will guarantee to fail when # Docker tries to run with --gpus all # # So, the correct check here is to query one of the missing piece of info like # GPU name, so that the command can fail accordingly nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 NVIDIA_SMI_STATUS=$? # Allowable exit statuses for nvidia-smi, see: https://github.com/NVIDIA/gpu-operator/issues/285 if [ "$NVIDIA_SMI_STATUS" -eq 0 ] || [ "$NVIDIA_SMI_STATUS" -eq 14 ]; then echo "INFO: Ignoring allowed status ${NVIDIA_SMI_STATUS}" else echo "ERROR: nvidia-smi exited with unresolved status ${NVIDIA_SMI_STATUS}" exit ${NVIDIA_SMI_STATUS} fi set -e ) ) } install_nvidia_driver_amzn2() { ( set -x pre_install_nvidia_driver_amzn2 install_nvidia_driver_common post_install_nvidia_driver_common ) } install_nvidia_driver_ubuntu20() { ( set -x install_nvidia_driver_common post_install_nvidia_driver_common ) } echo "== Installing nvidia driver ${DRIVER_FN} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_driver_amzn2 ;; ubuntu20.04) install_nvidia_driver_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Install container toolkit based on distribution echo "== Installing nvidia container toolkit for ${DISTRIBUTION} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_docker2_amzn2 ;; ubuntu20.04) install_nvidia_docker2_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Fix https://github.com/NVIDIA/nvidia-docker/issues/1648 on runners with # more than one GPUs. This just needs to be run once. The command fails # on subsequent runs and complains that the mode is already on, but that's # ok sudo nvidia-persistenced || true # This should show persistence mode ON nvidia-smi # check if the container-toolkit is correctly installed and CUDA is available inside a container docker run --rm -t --gpus=all public.ecr.aws/docker/library/python:3.13 nvidia-smi 2025-12-04T09:23:44.9397722Z retry_wait_seconds: 10 2025-12-04T09:23:44.9397996Z polling_interval_seconds: 1 2025-12-04T09:23:44.9398268Z warning_on_retry: true 2025-12-04T09:23:44.9398537Z continue_on_error: false 2025-12-04T09:23:44.9398786Z env: 2025-12-04T09:23:44.9398989Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:23:44.9399252Z HAS_NVIDIA_GPU: true 2025-12-04T09:23:44.9399565Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:23:44.9399922Z DRIVER_VERSION: 580.82.07 2025-12-04T09:23:44.9400176Z ##[endgroup] 2025-12-04T09:23:45.0777908Z == Installing nvidia driver NVIDIA-Linux-x86_64-580.82.07.run == 2025-12-04T09:23:45.0779021Z + pre_install_nvidia_driver_amzn2 2025-12-04T09:23:45.0782198Z + sudo yum remove -y nvidia-driver-latest-dkms 2025-12-04T09:23:45.7561952Z No match for argument: nvidia-driver-latest-dkms 2025-12-04T09:23:45.7562968Z No packages marked for removal. 2025-12-04T09:23:45.7630574Z Dependencies resolved. 2025-12-04T09:23:45.7641397Z Nothing to do. 2025-12-04T09:23:45.7641804Z Complete! 2025-12-04T09:23:45.8501298Z + install_nvidia_driver_common 2025-12-04T09:23:45.8505471Z + echo 'Before installing NVIDIA driver' 2025-12-04T09:23:45.8505947Z + lspci 2025-12-04T09:23:45.8507641Z Before installing NVIDIA driver 2025-12-04T09:23:45.9105877Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-12-04T09:23:45.9106404Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-12-04T09:23:45.9106976Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-12-04T09:23:45.9107513Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-12-04T09:23:45.9108009Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-12-04T09:23:45.9108557Z 00:05.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-12-04T09:23:45.9109141Z 00:1e.0 3D controller: NVIDIA Corporation GA102GL [A10G] (rev a1) 2025-12-04T09:23:45.9109766Z 00:1f.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-12-04T09:23:45.9110301Z + lsmod 2025-12-04T09:23:45.9162000Z Module Size Used by 2025-12-04T09:23:45.9162578Z nvidia_uvm 1925120 0 2025-12-04T09:23:45.9162859Z nvidia 14286848 1 nvidia_uvm 2025-12-04T09:23:45.9163155Z drm 602112 1 nvidia 2025-12-04T09:23:45.9163465Z drm_panel_orientation_quirks 32768 1 drm 2025-12-04T09:23:45.9163780Z backlight 24576 1 drm 2025-12-04T09:23:45.9164064Z i2c_core 110592 2 nvidia,drm 2025-12-04T09:23:45.9164357Z xt_conntrack 16384 1 2025-12-04T09:23:45.9164626Z nft_chain_nat 16384 3 2025-12-04T09:23:45.9164886Z xt_MASQUERADE 20480 1 2025-12-04T09:23:45.9165188Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-12-04T09:23:45.9165525Z nf_conntrack_netlink 57344 0 2025-12-04T09:23:45.9165929Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-12-04T09:23:45.9166522Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-12-04T09:23:45.9166849Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-12-04T09:23:45.9167226Z xfrm_user 57344 1 2025-12-04T09:23:45.9167499Z xfrm_algo 16384 1 xfrm_user 2025-12-04T09:23:45.9167808Z xt_addrtype 16384 2 2025-12-04T09:23:45.9168074Z nft_compat 20480 4 2025-12-04T09:23:45.9168383Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-12-04T09:23:45.9168810Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-12-04T09:23:45.9169191Z br_netfilter 36864 0 2025-12-04T09:23:45.9169471Z bridge 323584 1 br_netfilter 2025-12-04T09:23:45.9169769Z stp 16384 1 bridge 2025-12-04T09:23:45.9170057Z llc 16384 2 bridge,stp 2025-12-04T09:23:45.9170335Z overlay 167936 0 2025-12-04T09:23:45.9170593Z tls 139264 0 2025-12-04T09:23:45.9170858Z nls_ascii 16384 1 2025-12-04T09:23:45.9171116Z nls_cp437 20480 1 2025-12-04T09:23:45.9171361Z vfat 24576 1 2025-12-04T09:23:45.9171620Z fat 86016 1 vfat 2025-12-04T09:23:45.9171889Z sunrpc 700416 1 2025-12-04T09:23:45.9172148Z ghash_clmulni_intel 16384 0 2025-12-04T09:23:45.9172414Z i8042 45056 0 2025-12-04T09:23:45.9172673Z serio 28672 3 i8042 2025-12-04T09:23:45.9172934Z ena 184320 0 2025-12-04T09:23:45.9173187Z button 24576 0 2025-12-04T09:23:45.9173447Z sch_fq_codel 20480 17 2025-12-04T09:23:45.9173697Z fuse 184320 1 2025-12-04T09:23:45.9173946Z loop 36864 0 2025-12-04T09:23:45.9174199Z dm_mod 188416 0 2025-12-04T09:23:45.9174446Z configfs 57344 1 2025-12-04T09:23:45.9174700Z dmi_sysfs 20480 0 2025-12-04T09:23:45.9174956Z crc32_pclmul 16384 0 2025-12-04T09:23:45.9175226Z crc32c_intel 24576 0 2025-12-04T09:23:45.9175476Z efivarfs 24576 1 2025-12-04T09:23:45.9175729Z + modinfo nvidia 2025-12-04T09:23:45.9182462Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-12-04T09:23:45.9182949Z import_ns: DMA_BUF 2025-12-04T09:23:45.9183197Z alias: char-major-195-* 2025-12-04T09:23:45.9183468Z version: 580.82.07 2025-12-04T09:23:45.9183717Z supported: external 2025-12-04T09:23:45.9183961Z license: Dual MIT/GPL 2025-12-04T09:23:45.9184249Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-12-04T09:23:45.9184591Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-12-04T09:23:45.9184912Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-12-04T09:23:45.9185248Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-12-04T09:23:45.9185603Z alias: of:N*T*Cnvidia,tegra264-display 2025-12-04T09:23:45.9185952Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-12-04T09:23:45.9186302Z alias: of:N*T*Cnvidia,tegra234-display 2025-12-04T09:23:45.9186648Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-12-04T09:23:45.9187101Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-12-04T09:23:45.9187470Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-12-04T09:23:45.9187784Z depends: i2c-core,drm 2025-12-04T09:23:45.9188041Z retpoline: Y 2025-12-04T09:23:45.9188260Z name: nvidia 2025-12-04T09:23:45.9188621Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-12-04T09:23:45.9189102Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-12-04T09:23:45.9189557Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-12-04T09:23:45.9189976Z parm: NVreg_ResmanDebugLevel:int 2025-12-04T09:23:45.9190289Z parm: NVreg_RmLogonRC:int 2025-12-04T09:23:45.9190601Z parm: NVreg_ModifyDeviceFiles:int 2025-12-04T09:23:45.9191000Z parm: NVreg_DeviceFileUID:int 2025-12-04T09:23:45.9191316Z parm: NVreg_DeviceFileGID:int 2025-12-04T09:23:45.9191630Z parm: NVreg_DeviceFileMode:int 2025-12-04T09:23:45.9192002Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-12-04T09:23:45.9192402Z parm: NVreg_UsePageAttributeTable:int 2025-12-04T09:23:45.9192744Z parm: NVreg_EnablePCIeGen3:int 2025-12-04T09:23:45.9193049Z parm: NVreg_EnableMSI:int 2025-12-04T09:23:45.9193365Z parm: NVreg_EnableStreamMemOPs:int 2025-12-04T09:23:45.9193736Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-12-04T09:23:45.9194143Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-12-04T09:23:45.9194531Z parm: NVreg_EnableS0ixPowerManagement:int 2025-12-04T09:23:45.9194958Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-12-04T09:23:45.9195381Z parm: NVreg_DynamicPowerManagement:int 2025-12-04T09:23:45.9195809Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-12-04T09:23:45.9196239Z parm: NVreg_EnableGpuFirmware:int 2025-12-04T09:23:45.9196587Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-12-04T09:23:45.9196963Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-12-04T09:23:45.9197346Z parm: NVreg_EnableUserNUMAManagement:int 2025-12-04T09:23:45.9197695Z parm: NVreg_MemoryPoolSize:int 2025-12-04T09:23:45.9198025Z parm: NVreg_KMallocHeapMaxSize:int 2025-12-04T09:23:45.9198357Z parm: NVreg_VMallocHeapMaxSize:int 2025-12-04T09:23:45.9198686Z parm: NVreg_IgnoreMMIOCheck:int 2025-12-04T09:23:45.9199003Z parm: NVreg_NvLinkDisable:int 2025-12-04T09:23:45.9199351Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-12-04T09:23:45.9199722Z parm: NVreg_RegisterPCIDriver:int 2025-12-04T09:23:45.9200086Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-12-04T09:23:45.9200453Z parm: NVreg_EnableResizableBar:int 2025-12-04T09:23:45.9200802Z parm: NVreg_EnableDbgBreakpoint:int 2025-12-04T09:23:45.9201156Z parm: NVreg_EnableNonblockingOpen:int 2025-12-04T09:23:45.9201522Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-12-04T09:23:45.9201863Z parm: NVreg_RegistryDwords:charp 2025-12-04T09:23:45.9202214Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-12-04T09:23:45.9202551Z parm: NVreg_RmMsg:charp 2025-12-04T09:23:45.9202840Z parm: NVreg_GpuBlacklist:charp 2025-12-04T09:23:45.9203175Z parm: NVreg_TemporaryFilePath:charp 2025-12-04T09:23:45.9203507Z parm: NVreg_ExcludedGpus:charp 2025-12-04T09:23:45.9203824Z parm: NVreg_DmaRemapPeerMmio:int 2025-12-04T09:23:45.9204160Z parm: NVreg_RmNvlinkBandwidth:charp 2025-12-04T09:23:45.9204529Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-12-04T09:23:45.9204890Z parm: NVreg_ImexChannelCount:int 2025-12-04T09:23:45.9205223Z parm: NVreg_CreateImexChannel0:int 2025-12-04T09:23:45.9205587Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-12-04T09:23:45.9205940Z parm: rm_firmware_active:charp 2025-12-04T09:23:45.9206332Z + HAS_NVIDIA_DRIVER=0 2025-12-04T09:23:45.9206587Z ++ command -v nvidia-smi 2025-12-04T09:23:45.9206852Z + '[' -x /usr/bin/nvidia-smi ']' 2025-12-04T09:23:45.9207228Z + set +e 2025-12-04T09:23:45.9207574Z ++ nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0 2025-12-04T09:23:47.6466249Z + INSTALLED_DRIVER_VERSION=580.82.07 2025-12-04T09:23:47.6466593Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:23:47.6466842Z + '[' 0 -ne 0 ']' 2025-12-04T09:23:47.6467072Z + '[' 580.82.07 '!=' 580.82.07 ']' 2025-12-04T09:23:47.6467340Z + HAS_NVIDIA_DRIVER=1 2025-12-04T09:23:47.6467780Z + echo 'NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation' 2025-12-04T09:23:47.6468246Z + set -e 2025-12-04T09:23:47.6468442Z + '[' 1 -eq 0 ']' 2025-12-04T09:23:47.6468833Z NVIDIA driver (580.82.07) has already been installed. Skipping NVIDIA driver installation 2025-12-04T09:23:47.6470055Z + post_install_nvidia_driver_common 2025-12-04T09:23:47.6473380Z + sudo modprobe nvidia 2025-12-04T09:23:47.8046957Z + echo 'After installing NVIDIA driver' 2025-12-04T09:23:47.8047347Z + lspci 2025-12-04T09:23:47.8047571Z After installing NVIDIA driver 2025-12-04T09:23:47.8170911Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-12-04T09:23:47.8171422Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-12-04T09:23:47.8171993Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-12-04T09:23:47.8172540Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-12-04T09:23:47.8173034Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-12-04T09:23:47.8173581Z 00:05.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-12-04T09:23:47.8174082Z 00:1e.0 3D controller: NVIDIA Corporation GA102GL [A10G] (rev a1) 2025-12-04T09:23:47.8174584Z 00:1f.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-12-04T09:23:47.8175003Z + lsmod 2025-12-04T09:23:47.8214252Z Module Size Used by 2025-12-04T09:23:47.8214565Z nvidia_uvm 1925120 0 2025-12-04T09:23:47.8215222Z nvidia 14286848 1 nvidia_uvm 2025-12-04T09:23:47.8215576Z drm 602112 1 nvidia 2025-12-04T09:23:47.8215890Z drm_panel_orientation_quirks 32768 1 drm 2025-12-04T09:23:47.8216208Z backlight 24576 1 drm 2025-12-04T09:23:47.8216502Z i2c_core 110592 2 nvidia,drm 2025-12-04T09:23:47.8216797Z xt_conntrack 16384 1 2025-12-04T09:23:47.8217191Z nft_chain_nat 16384 3 2025-12-04T09:23:47.8229268Z xt_MASQUERADE 20480 1 2025-12-04T09:23:47.8229714Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-12-04T09:23:47.8230166Z nf_conntrack_netlink 57344 0 2025-12-04T09:23:47.8230752Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-12-04T09:23:47.8231245Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-12-04T09:23:47.8231583Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-12-04T09:23:47.8231877Z xfrm_user 57344 1 2025-12-04T09:23:47.8232151Z xfrm_algo 16384 1 xfrm_user 2025-12-04T09:23:47.8232450Z xt_addrtype 16384 2 2025-12-04T09:23:47.8232707Z nft_compat 20480 4 2025-12-04T09:23:47.8233026Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-12-04T09:23:47.8233455Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-12-04T09:23:47.8233835Z br_netfilter 36864 0 2025-12-04T09:23:47.8234124Z bridge 323584 1 br_netfilter 2025-12-04T09:23:47.8234426Z stp 16384 1 bridge 2025-12-04T09:23:47.8234716Z llc 16384 2 bridge,stp 2025-12-04T09:23:47.8234991Z overlay 167936 0 2025-12-04T09:23:47.8235547Z tls 139264 0 2025-12-04T09:23:47.8235823Z nls_ascii 16384 1 2025-12-04T09:23:47.8236071Z nls_cp437 20480 1 2025-12-04T09:23:47.8236653Z vfat 24576 1 2025-12-04T09:23:47.8236923Z fat 86016 1 vfat 2025-12-04T09:23:47.8237290Z sunrpc 700416 1 2025-12-04T09:23:47.8237654Z ghash_clmulni_intel 16384 0 2025-12-04T09:23:47.8238066Z i8042 45056 0 2025-12-04T09:23:47.8238404Z serio 28672 3 i8042 2025-12-04T09:23:47.8238754Z ena 184320 0 2025-12-04T09:23:47.8239070Z button 24576 0 2025-12-04T09:23:47.8239329Z sch_fq_codel 20480 17 2025-12-04T09:23:47.8239578Z fuse 184320 1 2025-12-04T09:23:47.8239826Z loop 36864 0 2025-12-04T09:23:47.8240075Z dm_mod 188416 0 2025-12-04T09:23:47.8240318Z configfs 57344 1 2025-12-04T09:23:47.8240570Z dmi_sysfs 20480 0 2025-12-04T09:23:47.8241013Z crc32_pclmul 16384 0 2025-12-04T09:23:47.8241260Z crc32c_intel 24576 0 2025-12-04T09:23:47.8241516Z efivarfs 24576 1 2025-12-04T09:23:47.8241764Z + modinfo nvidia 2025-12-04T09:23:47.8242191Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-12-04T09:23:47.8242630Z import_ns: DMA_BUF 2025-12-04T09:23:47.8242923Z alias: char-major-195-* 2025-12-04T09:23:47.8243296Z version: 580.82.07 2025-12-04T09:23:47.8243625Z supported: external 2025-12-04T09:23:47.8243956Z license: Dual MIT/GPL 2025-12-04T09:23:47.8244331Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-12-04T09:23:47.8244680Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-12-04T09:23:47.8245000Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-12-04T09:23:47.8245339Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-12-04T09:23:47.8245695Z alias: of:N*T*Cnvidia,tegra264-display 2025-12-04T09:23:47.8246050Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-12-04T09:23:47.8246395Z alias: of:N*T*Cnvidia,tegra234-display 2025-12-04T09:23:47.8246740Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-12-04T09:23:47.8247276Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-12-04T09:23:47.8247822Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-12-04T09:23:47.8248310Z depends: i2c-core,drm 2025-12-04T09:23:47.8248698Z retpoline: Y 2025-12-04T09:23:47.8249009Z name: nvidia 2025-12-04T09:23:47.8249500Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-12-04T09:23:47.8250179Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-12-04T09:23:47.8250699Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-12-04T09:23:47.8251173Z parm: NVreg_ResmanDebugLevel:int 2025-12-04T09:23:47.8251515Z parm: NVreg_RmLogonRC:int 2025-12-04T09:23:47.8251849Z parm: NVreg_ModifyDeviceFiles:int 2025-12-04T09:23:47.8252195Z parm: NVreg_DeviceFileUID:int 2025-12-04T09:23:47.8252525Z parm: NVreg_DeviceFileGID:int 2025-12-04T09:23:47.8252868Z parm: NVreg_DeviceFileMode:int 2025-12-04T09:23:47.8253260Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-12-04T09:23:47.8253699Z parm: NVreg_UsePageAttributeTable:int 2025-12-04T09:23:47.8254039Z parm: NVreg_EnablePCIeGen3:int 2025-12-04T09:23:47.8254339Z parm: NVreg_EnableMSI:int 2025-12-04T09:23:47.8254650Z parm: NVreg_EnableStreamMemOPs:int 2025-12-04T09:23:47.8255020Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-12-04T09:23:47.8255423Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-12-04T09:23:47.8255813Z parm: NVreg_EnableS0ixPowerManagement:int 2025-12-04T09:23:47.8256238Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-12-04T09:23:47.8256660Z parm: NVreg_DynamicPowerManagement:int 2025-12-04T09:23:47.8257087Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-12-04T09:23:47.8257654Z parm: NVreg_EnableGpuFirmware:int 2025-12-04T09:23:47.8258000Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-12-04T09:23:47.8258373Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-12-04T09:23:47.8258759Z parm: NVreg_EnableUserNUMAManagement:int 2025-12-04T09:23:47.8259102Z parm: NVreg_MemoryPoolSize:int 2025-12-04T09:23:47.8259423Z parm: NVreg_KMallocHeapMaxSize:int 2025-12-04T09:23:47.8259760Z parm: NVreg_VMallocHeapMaxSize:int 2025-12-04T09:23:47.8260091Z parm: NVreg_IgnoreMMIOCheck:int 2025-12-04T09:23:47.8260408Z parm: NVreg_NvLinkDisable:int 2025-12-04T09:23:47.8260757Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-12-04T09:23:47.8261125Z parm: NVreg_RegisterPCIDriver:int 2025-12-04T09:23:47.8261484Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-12-04T09:23:47.8261945Z parm: NVreg_EnableResizableBar:int 2025-12-04T09:23:47.8262285Z parm: NVreg_EnableDbgBreakpoint:int 2025-12-04T09:23:47.8262643Z parm: NVreg_EnableNonblockingOpen:int 2025-12-04T09:23:47.8263005Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-12-04T09:23:47.8263360Z parm: NVreg_RegistryDwords:charp 2025-12-04T09:23:47.8263708Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-12-04T09:23:47.8264039Z parm: NVreg_RmMsg:charp 2025-12-04T09:23:47.8264338Z parm: NVreg_GpuBlacklist:charp 2025-12-04T09:23:47.8264664Z parm: NVreg_TemporaryFilePath:charp 2025-12-04T09:23:47.8264995Z parm: NVreg_ExcludedGpus:charp 2025-12-04T09:23:47.8265311Z parm: NVreg_DmaRemapPeerMmio:int 2025-12-04T09:23:47.8265643Z parm: NVreg_RmNvlinkBandwidth:charp 2025-12-04T09:23:47.8266005Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-12-04T09:23:47.8266357Z parm: NVreg_ImexChannelCount:int 2025-12-04T09:23:47.8266696Z parm: NVreg_CreateImexChannel0:int 2025-12-04T09:23:47.8267052Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-12-04T09:23:47.8267395Z parm: rm_firmware_active:charp 2025-12-04T09:23:47.8267685Z + set +e 2025-12-04T09:23:47.8267880Z + nvidia-smi 2025-12-04T09:23:49.2769336Z Thu Dec 4 09:23:49 2025 2025-12-04T09:23:49.2769719Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:23:49.2770222Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:23:49.2770720Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:23:49.2771221Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:23:49.2771753Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:23:49.2772215Z | | | MIG M. | 2025-12-04T09:23:49.2772566Z |=========================================+========================+======================| 2025-12-04T09:23:49.2858267Z | 0 NVIDIA A10G Off | 00000000:00:1E.0 Off | 0 | 2025-12-04T09:23:49.2858837Z | 0% 27C P0 64W / 300W | 0MiB / 23028MiB | 4% Default | 2025-12-04T09:23:49.2859305Z | | | N/A | 2025-12-04T09:23:49.2859709Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:23:49.2860024Z 2025-12-04T09:23:49.2860246Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:23:49.2860686Z | Processes: | 2025-12-04T09:23:49.2861136Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:23:49.2861562Z | ID ID Usage | 2025-12-04T09:23:49.2862186Z |=========================================================================================| 2025-12-04T09:23:49.2864058Z | No running processes found | 2025-12-04T09:23:49.2864665Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:23:49.7095182Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-12-04T09:23:51.1613935Z NVIDIA A10G 2025-12-04T09:23:51.4323940Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:23:51.4324221Z + '[' 0 -eq 0 ']' 2025-12-04T09:23:51.4324493Z + echo 'INFO: Ignoring allowed status 0' 2025-12-04T09:23:51.4324792Z + set -e 2025-12-04T09:23:51.4325003Z INFO: Ignoring allowed status 0 2025-12-04T09:23:51.4333502Z == Installing nvidia container toolkit for amzn2023 == 2025-12-04T09:23:51.4338416Z + sudo yum install -y yum-utils 2025-12-04T09:23:51.9060844Z Last metadata expiration check: 0:08:47 ago on Thu Dec 4 09:15:04 2025. 2025-12-04T09:23:51.9329962Z Package dnf-utils-4.3.0-13.amzn2023.0.5.noarch is already installed. 2025-12-04T09:23:51.9913021Z Dependencies resolved. 2025-12-04T09:23:52.0205119Z Nothing to do. 2025-12-04T09:23:52.0205367Z Complete! 2025-12-04T09:23:52.1254846Z + [[ amzn2023 == \a\m\z\n\2\0\2\3 ]] 2025-12-04T09:23:52.1256125Z + YUM_REPO_URL=https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:23:52.1257127Z + sudo yum-config-manager --add-repo https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:23:52.4459032Z Adding repo from: https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:23:52.4998986Z + sudo yum install -y nvidia-container-toolkit-1.17.8 libnvidia-container-tools-1.17.8 libnvidia-container1-1.17.8 nvidia-container-toolkit-base-1.17.8 2025-12-04T09:23:53.0190674Z nvidia-container-toolkit 19 kB/s | 833 B 00:00 2025-12-04T09:23:53.1028413Z Dependencies resolved. 2025-12-04T09:23:53.1321229Z ================================================================================ 2025-12-04T09:23:53.1322402Z Package Arch Version Repository Size 2025-12-04T09:23:53.1322842Z ================================================================================ 2025-12-04T09:23:53.1323155Z Downgrading: 2025-12-04T09:23:53.1323532Z libnvidia-container-tools x86_64 1.17.8-1 nvidia-container-toolkit 40 k 2025-12-04T09:23:53.1324114Z libnvidia-container1 x86_64 1.17.8-1 nvidia-container-toolkit 1.0 M 2025-12-04T09:23:53.1324735Z nvidia-container-toolkit x86_64 1.17.8-1 nvidia-container-toolkit 1.2 M 2025-12-04T09:23:53.1325473Z nvidia-container-toolkit-base x86_64 1.17.8-1 nvidia-container-toolkit 5.8 M 2025-12-04T09:23:53.1325870Z 2025-12-04T09:23:53.1325963Z Transaction Summary 2025-12-04T09:23:53.1326281Z ================================================================================ 2025-12-04T09:23:53.1326656Z Downgrade 4 Packages 2025-12-04T09:23:53.1326814Z 2025-12-04T09:23:53.1326918Z Total download size: 8.0 M 2025-12-04T09:23:53.1327297Z Downloading Packages: 2025-12-04T09:23:53.1779508Z (1/4): libnvidia-container-tools-1.17.8-1.x86_6 922 kB/s | 40 kB 00:00 2025-12-04T09:23:53.2176742Z (2/4): libnvidia-container1-1.17.8-1.x86_64.rpm 12 MB/s | 1.0 MB 00:00 2025-12-04T09:23:53.2816527Z (3/4): nvidia-container-toolkit-1.17.8-1.x86_64 8.4 MB/s | 1.2 MB 00:00 2025-12-04T09:23:53.4186040Z (4/4): nvidia-container-toolkit-base-1.17.8-1.x 24 MB/s | 5.8 MB 00:00 2025-12-04T09:23:53.4197665Z -------------------------------------------------------------------------------- 2025-12-04T09:23:53.4200837Z Total 28 MB/s | 8.0 MB 00:00 2025-12-04T09:23:53.4203887Z Running transaction check 2025-12-04T09:23:53.4321837Z Transaction check succeeded. 2025-12-04T09:23:53.4322421Z Running transaction test 2025-12-04T09:23:53.4845732Z Transaction test succeeded. 2025-12-04T09:23:53.4846765Z Running transaction 2025-12-04T09:23:54.3411725Z Preparing : 1/1 2025-12-04T09:23:54.4704823Z Downgrading : nvidia-container-toolkit-base-1.17.8-1.x86_64 1/8 2025-12-04T09:23:54.4966763Z Downgrading : libnvidia-container1-1.17.8-1.x86_64 2/8 2025-12-04T09:23:54.5680550Z Running scriptlet: libnvidia-container1-1.17.8-1.x86_64 2/8 2025-12-04T09:23:54.6964854Z Downgrading : libnvidia-container-tools-1.17.8-1.x86_64 3/8 2025-12-04T09:23:54.7193684Z Downgrading : nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-12-04T09:23:54.8116492Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-12-04T09:23:54.8194186Z Running scriptlet: nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:23:54.8194822Z Cleanup : nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:23:54.8517663Z Running scriptlet: nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:23:54.8582207Z Running scriptlet: libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:23:54.8582956Z Cleanup : libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:23:54.8943829Z Running scriptlet: libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:23:54.9027521Z Running scriptlet: libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:23:54.9028107Z Cleanup : libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:23:54.9452834Z Running scriptlet: libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:23:54.9530154Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:23:54.9530927Z Cleanup : nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:23:54.9876312Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:23:55.0486556Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 8/8 2025-12-04T09:24:20.0337573Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:24:20.0338187Z Verifying : libnvidia-container-tools-1.17.8-1.x86_64 1/8 2025-12-04T09:24:20.0338736Z Verifying : libnvidia-container-tools-1.18.1-1.x86_64 2/8 2025-12-04T09:24:20.0339266Z Verifying : libnvidia-container1-1.17.8-1.x86_64 3/8 2025-12-04T09:24:20.0339795Z Verifying : libnvidia-container1-1.18.1-1.x86_64 4/8 2025-12-04T09:24:20.0340325Z Verifying : nvidia-container-toolkit-1.17.8-1.x86_64 5/8 2025-12-04T09:24:20.0340879Z Verifying : nvidia-container-toolkit-1.18.1-1.x86_64 6/8 2025-12-04T09:24:20.0341437Z Verifying : nvidia-container-toolkit-base-1.17.8-1.x86_64 7/8 2025-12-04T09:24:20.1866558Z Verifying : nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8================================================================================ 2025-12-04T09:24:20.1867104Z WARNING: 2025-12-04T09:24:20.1867387Z A newer release of "Amazon Linux" is available. 2025-12-04T09:24:20.1867631Z 2025-12-04T09:24:20.1867726Z Available Versions: 2025-12-04T09:24:20.1867880Z 2025-12-04T09:24:20.1867979Z Version 2023.9.20250929: 2025-12-04T09:24:20.1868297Z Run the following command to upgrade to 2023.9.20250929: 2025-12-04T09:24:20.1868563Z 2025-12-04T09:24:20.1868688Z dnf upgrade --releasever=2023.9.20250929 2025-12-04T09:24:20.1868907Z 2025-12-04T09:24:20.1869002Z Release notes: 2025-12-04T09:24:20.1869430Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20250929.html 2025-12-04T09:24:20.1869818Z 2025-12-04T09:24:20.1870247Z Version 2023.9.20251014: 2025-12-04T09:24:20.1870575Z Run the following command to upgrade to 2023.9.20251014: 2025-12-04T09:24:20.1870829Z 2025-12-04T09:24:20.1870964Z dnf upgrade --releasever=2023.9.20251014 2025-12-04T09:24:20.1871180Z 2025-12-04T09:24:20.1871268Z Release notes: 2025-12-04T09:24:20.1871678Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251014.html 2025-12-04T09:24:20.1872059Z 2025-12-04T09:24:20.1872152Z Version 2023.9.20251020: 2025-12-04T09:24:20.1872476Z Run the following command to upgrade to 2023.9.20251020: 2025-12-04T09:24:20.1872731Z 2025-12-04T09:24:20.1872849Z dnf upgrade --releasever=2023.9.20251020 2025-12-04T09:24:20.1873068Z 2025-12-04T09:24:20.1873155Z Release notes: 2025-12-04T09:24:20.1873551Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251020.html 2025-12-04T09:24:20.1874097Z 2025-12-04T09:24:20.1874187Z Version 2023.9.20251027: 2025-12-04T09:24:20.1874506Z Run the following command to upgrade to 2023.9.20251027: 2025-12-04T09:24:20.1874774Z 2025-12-04T09:24:20.1874889Z dnf upgrade --releasever=2023.9.20251027 2025-12-04T09:24:20.1875099Z 2025-12-04T09:24:20.1875186Z Release notes: 2025-12-04T09:24:20.1875580Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251027.html 2025-12-04T09:24:20.1875952Z 2025-12-04T09:24:20.1876039Z Version 2023.9.20251105: 2025-12-04T09:24:20.1876351Z Run the following command to upgrade to 2023.9.20251105: 2025-12-04T09:24:20.1876600Z 2025-12-04T09:24:20.1876723Z dnf upgrade --releasever=2023.9.20251105 2025-12-04T09:24:20.1876933Z 2025-12-04T09:24:20.1877016Z Release notes: 2025-12-04T09:24:20.1877411Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251105.html 2025-12-04T09:24:20.1877785Z 2025-12-04T09:24:20.1877880Z Version 2023.9.20251110: 2025-12-04T09:24:20.1878184Z Run the following command to upgrade to 2023.9.20251110: 2025-12-04T09:24:20.1878447Z 2025-12-04T09:24:20.1878564Z dnf upgrade --releasever=2023.9.20251110 2025-12-04T09:24:20.1878785Z 2025-12-04T09:24:20.1878868Z Release notes: 2025-12-04T09:24:20.1879268Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251110.html 2025-12-04T09:24:20.1879634Z 2025-12-04T09:24:20.1879724Z Version 2023.9.20251117: 2025-12-04T09:24:20.1880038Z Run the following command to upgrade to 2023.9.20251117: 2025-12-04T09:24:20.1880344Z 2025-12-04T09:24:20.1880460Z dnf upgrade --releasever=2023.9.20251117 2025-12-04T09:24:20.1880671Z 2025-12-04T09:24:20.1880760Z Release notes: 2025-12-04T09:24:20.1881147Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251117.html 2025-12-04T09:24:20.1881519Z 2025-12-04T09:24:20.1881625Z ================================================================================ 2025-12-04T09:24:20.2459256Z 2025-12-04T09:24:20.2459778Z 2025-12-04T09:24:20.2460174Z Downgraded: 2025-12-04T09:24:20.2460895Z libnvidia-container-tools-1.17.8-1.x86_64 2025-12-04T09:24:20.2461984Z libnvidia-container1-1.17.8-1.x86_64 2025-12-04T09:24:20.2463053Z nvidia-container-toolkit-1.17.8-1.x86_64 2025-12-04T09:24:20.2464167Z nvidia-container-toolkit-base-1.17.8-1.x86_64 2025-12-04T09:24:20.2464858Z 2025-12-04T09:24:20.2465015Z Complete! 2025-12-04T09:24:20.3234993Z + sudo systemctl restart docker 2025-12-04T09:24:26.7267690Z Thu Dec 4 09:24:26 2025 2025-12-04T09:24:26.7268098Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:24:26.7268597Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:24:26.7269109Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:24:26.7269943Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:24:26.7270478Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:24:26.7270919Z | | | MIG M. | 2025-12-04T09:24:26.7271250Z |=========================================+========================+======================| 2025-12-04T09:24:26.7365218Z | 0 NVIDIA A10G On | 00000000:00:1E.0 Off | 0 | 2025-12-04T09:24:26.7366544Z | 0% 27C P0 60W / 300W | 0MiB / 23028MiB | 4% Default | 2025-12-04T09:24:26.7367449Z | | | N/A | 2025-12-04T09:24:26.7367906Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:24:26.7368405Z 2025-12-04T09:24:26.7368586Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:24:26.7369142Z | Processes: | 2025-12-04T09:24:26.7369752Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:24:26.7370170Z | ID ID Usage | 2025-12-04T09:24:26.7370517Z |=========================================================================================| 2025-12-04T09:24:26.7370937Z | No running processes found | 2025-12-04T09:24:26.7371404Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:24:26.9118777Z Unable to find image 'public.ecr.aws/docker/library/python:3.13' locally 2025-12-04T09:24:27.1331252Z 3.13: Pulling from docker/library/python 2025-12-04T09:24:27.2283785Z 53c88f1dfeb7: Pulling fs layer 2025-12-04T09:24:27.2284143Z eae668646f44: Pulling fs layer 2025-12-04T09:24:27.2284419Z ff2e6e687b6c: Pulling fs layer 2025-12-04T09:24:27.2284690Z 7c40a3faff76: Pulling fs layer 2025-12-04T09:24:27.2284953Z 967a3b1c8fef: Pulling fs layer 2025-12-04T09:24:27.2285221Z a64e1a44f22a: Pulling fs layer 2025-12-04T09:24:27.2285601Z 52655f8a5bcc: Pulling fs layer 2025-12-04T09:24:27.2285861Z 967a3b1c8fef: Waiting 2025-12-04T09:24:27.2286090Z a64e1a44f22a: Waiting 2025-12-04T09:24:27.2286315Z 52655f8a5bcc: Waiting 2025-12-04T09:24:27.2286533Z 7c40a3faff76: Waiting 2025-12-04T09:24:27.3392663Z eae668646f44: Verifying Checksum 2025-12-04T09:24:27.3393129Z eae668646f44: Download complete 2025-12-04T09:24:27.4217024Z 53c88f1dfeb7: Verifying Checksum 2025-12-04T09:24:27.4217504Z 53c88f1dfeb7: Download complete 2025-12-04T09:24:27.5248664Z 967a3b1c8fef: Verifying Checksum 2025-12-04T09:24:27.5249074Z 967a3b1c8fef: Download complete 2025-12-04T09:24:27.6185625Z ff2e6e687b6c: Verifying Checksum 2025-12-04T09:24:27.6185997Z ff2e6e687b6c: Download complete 2025-12-04T09:24:27.6665972Z 52655f8a5bcc: Download complete 2025-12-04T09:24:27.6866465Z a64e1a44f22a: Verifying Checksum 2025-12-04T09:24:27.6866907Z a64e1a44f22a: Download complete 2025-12-04T09:24:28.0295946Z 7c40a3faff76: Verifying Checksum 2025-12-04T09:24:28.0296410Z 7c40a3faff76: Download complete 2025-12-04T09:24:29.2192674Z 53c88f1dfeb7: Pull complete 2025-12-04T09:24:29.9413539Z eae668646f44: Pull complete 2025-12-04T09:24:32.4790746Z ff2e6e687b6c: Pull complete 2025-12-04T09:24:39.2959616Z 7c40a3faff76: Pull complete 2025-12-04T09:24:39.5883455Z 967a3b1c8fef: Pull complete 2025-12-04T09:24:40.3653599Z a64e1a44f22a: Pull complete 2025-12-04T09:24:40.3888508Z 52655f8a5bcc: Pull complete 2025-12-04T09:24:40.4023116Z Digest: sha256:3f986299a7b8b44b0d8cf9bda2b22361ce5c3058ef5d7cb17fb7452506680ab0 2025-12-04T09:24:40.4065969Z Status: Downloaded newer image for public.ecr.aws/docker/library/python:3.13 2025-12-04T09:24:47.5670406Z Thu Dec 4 09:24:47 2025 2025-12-04T09:24:47.5671240Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:24:47.5671849Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:24:47.5672341Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:24:47.5672913Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:24:47.5673520Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:24:47.5673961Z | | | MIG M. | 2025-12-04T09:24:47.5674372Z |=========================================+========================+======================| 2025-12-04T09:24:47.5819226Z | 0 NVIDIA A10G On | 00000000:00:1E.0 Off | 0 | 2025-12-04T09:24:47.5819771Z | 0% 25C P8 11W / 300W | 0MiB / 23028MiB | 0% Default | 2025-12-04T09:24:47.5820181Z | | | N/A | 2025-12-04T09:24:47.5820579Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:24:47.5823319Z 2025-12-04T09:24:47.5823652Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:24:47.5824094Z | Processes: | 2025-12-04T09:24:47.5824542Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:24:47.5824960Z | ID ID Usage | 2025-12-04T09:24:47.5825306Z |=========================================================================================| 2025-12-04T09:24:47.5829875Z | No running processes found | 2025-12-04T09:24:47.5830393Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:24:49.0784951Z Command completed after 1 attempt(s). 2025-12-04T09:24:49.0905889Z Prepare all required actions 2025-12-04T09:24:49.0935035Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-12-04T09:24:49.0935632Z with: 2025-12-04T09:24:49.0936255Z github-token: *** 2025-12-04T09:24:49.0936485Z env: 2025-12-04T09:24:49.0936697Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:24:49.0936964Z HAS_NVIDIA_GPU: true 2025-12-04T09:24:49.0937277Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:24:49.0937628Z ##[endgroup] 2025-12-04T09:24:49.0952324Z ##[group]Run set -eux 2025-12-04T09:24:49.0952575Z set -eux 2025-12-04T09:24:49.0953008Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T09:24:49.0968274Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:24:49.0968630Z env: 2025-12-04T09:24:49.0968840Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:24:49.0969105Z HAS_NVIDIA_GPU: true 2025-12-04T09:24:49.0969444Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:24:49.0969917Z GITHUB_TOKEN: *** 2025-12-04T09:24:49.0970141Z ##[endgroup] 2025-12-04T09:24:49.1012745Z + python3 .github/scripts/get_workflow_job_id.py 19922826259 i-016559b86632fc7da 2025-12-04T09:24:50.7371930Z Setting output job-id=57118183167 2025-12-04T09:24:50.7454561Z Setting output job-name=linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:24:50.7495606Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:24:50.7496419Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:24:50.7497544Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-12-04T09:24:50.7498399Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:24:50.7508477Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:24:50.7508924Z env: 2025-12-04T09:24:50.7509217Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:24:50.7509639Z HAS_NVIDIA_GPU: true 2025-12-04T09:24:50.7510039Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:24:50.7510534Z JOB_ID: 57118183167 2025-12-04T09:24:50.7511357Z JOB_NAME: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:24:50.7512188Z WORKFLOW_NAME: periodic 2025-12-04T09:24:50.7512614Z WORKFLOW_RUN_ID: 19922826259 2025-12-04T09:24:50.7513155Z MONITOR_LOG_INTERVAL: 5 2025-12-04T09:24:50.7513495Z MONITOR_DATA_COLLECT_INTERVAL: 1 2025-12-04T09:24:50.7513981Z ##[endgroup] 2025-12-04T09:24:51.0442066Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:24:51.4467952Z Collecting psutil==5.9.8 2025-12-04T09:24:51.4635202Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-12-04T09:24:51.5408351Z Collecting dataclasses_json==0.6.7 2025-12-04T09:24:51.5439615Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-12-04T09:24:51.5730075Z Collecting nvidia-ml-py==11.525.84 2025-12-04T09:24:51.5764080Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-12-04T09:24:51.7079474Z Collecting marshmallow<4.0.0,>=3.18.0 2025-12-04T09:24:51.7110999Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-12-04T09:24:51.7354781Z Collecting typing-inspect<1,>=0.4.0 2025-12-04T09:24:51.7400309Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-12-04T09:24:51.7991414Z Collecting packaging>=17.0 2025-12-04T09:24:51.8022991Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-12-04T09:24:51.8578276Z Collecting typing-extensions>=3.7.4 2025-12-04T09:24:51.8610099Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-12-04T09:24:51.8815608Z Collecting mypy-extensions>=0.3.0 2025-12-04T09:24:51.8847289Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-12-04T09:24:51.9768898Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-12-04T09:24:52.2520770Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-12-04T09:24:52.4495763Z Prepare all required actions 2025-12-04T09:24:52.4496314Z Getting action download info 2025-12-04T09:24:52.6085030Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:24:52.8488788Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-12-04T09:24:53.2197553Z ##[group]Run ./.github/actions/download-build-artifacts 2025-12-04T09:24:53.2197897Z with: 2025-12-04T09:24:53.2198825Z name: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck 2025-12-04T09:24:53.2199183Z s3-bucket: gha-artifacts 2025-12-04T09:24:53.2199437Z env: 2025-12-04T09:24:53.2199640Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:24:53.2199897Z HAS_NVIDIA_GPU: true 2025-12-04T09:24:53.2200211Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:24:53.2200552Z ##[endgroup] 2025-12-04T09:24:53.2231001Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:24:53.2231370Z with: 2025-12-04T09:24:53.2231641Z name: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck 2025-12-04T09:24:53.2232015Z s3-bucket: gha-artifacts 2025-12-04T09:24:53.2232274Z region: us-east-1 2025-12-04T09:24:53.2232491Z env: 2025-12-04T09:24:53.2232690Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:24:53.2232948Z HAS_NVIDIA_GPU: true 2025-12-04T09:24:53.2233259Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:24:53.2233598Z ##[endgroup] 2025-12-04T09:24:53.6978327Z (node:59447) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:24:53.6978808Z 2025-12-04T09:24:53.6979003Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:24:53.6979529Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:24:53.6980080Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:24:53.9838622Z Found 1 objects with prefix pytorch/pytorch/19922826259/linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck/ 2025-12-04T09:24:53.9839396Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:25:05.9693484Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:25:05.9699442Z Artifact download has finished successfully 2025-12-04T09:25:06.0073041Z ##[group]Run unzip -o artifacts.zip 2025-12-04T09:25:06.0073378Z unzip -o artifacts.zip 2025-12-04T09:25:06.0167457Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:06.0167820Z env: 2025-12-04T09:25:06.0168056Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:06.0168310Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:06.0168619Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:06.0168950Z ##[endgroup] 2025-12-04T09:25:06.0248930Z Archive: artifacts.zip 2025-12-04T09:25:06.0250764Z creating: dist/ 2025-12-04T09:25:08.0931851Z inflating: dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:25:08.1068412Z inflating: dist/.ninja_log 2025-12-04T09:25:08.1069201Z creating: build/custom_test_artifacts/ 2025-12-04T09:25:08.1070201Z creating: build/custom_test_artifacts/custom-op-build/ 2025-12-04T09:25:08.1071446Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-12-04T09:25:08.1072946Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:25:08.1079764Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:25:08.1080623Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-12-04T09:25:08.1081502Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:25:08.1082417Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:25:08.1083456Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:25:08.1085633Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:25:08.1087178Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:25:08.1088414Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:25:08.1089364Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:25:08.1090266Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:25:08.1092941Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:25:08.1094252Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:25:08.1095587Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:25:08.1097683Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:25:08.1099967Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:25:08.1101007Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:25:08.1101741Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:25:08.1162796Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:25:08.1222743Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:25:08.1223997Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:25:08.1289361Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:25:08.1290843Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:25:08.1292134Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:25:08.1293398Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:25:08.1294611Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:25:08.1295895Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:25:08.1297189Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:25:08.1298453Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:25:08.1299725Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:25:08.1301168Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:25:08.1302340Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:25:08.1303448Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:25:08.1305048Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:25:08.1306334Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:25:08.1309416Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:25:08.1385022Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:25:08.1386042Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:25:08.1462289Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:25:08.1463246Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:25:08.1464043Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:25:08.1464869Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-12-04T09:25:08.1465713Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-12-04T09:25:08.1466663Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-12-04T09:25:08.1467737Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-12-04T09:25:08.1468761Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-12-04T09:25:08.1469646Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-12-04T09:25:08.1470633Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-12-04T09:25:08.1471625Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-12-04T09:25:08.1472574Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-12-04T09:25:08.1473528Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-12-04T09:25:08.1474285Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-12-04T09:25:08.1494653Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-12-04T09:25:08.1698900Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-12-04T09:25:08.1699826Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-12-04T09:25:08.1700836Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-12-04T09:25:08.1701987Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-12-04T09:25:08.1703071Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-12-04T09:25:08.1703995Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-12-04T09:25:08.1705039Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-12-04T09:25:08.1706100Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-12-04T09:25:08.1707155Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-12-04T09:25:08.1708024Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-12-04T09:25:08.1708845Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-12-04T09:25:08.1728994Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-12-04T09:25:08.1812501Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-12-04T09:25:08.1813970Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:25:08.1814993Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:25:08.1815909Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-12-04T09:25:08.1816685Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-12-04T09:25:08.1818835Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-12-04T09:25:08.1819681Z inflating: build/custom_test_artifacts/custom-op-build/detect_cuda_version.cc 2025-12-04T09:25:08.1822216Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-12-04T09:25:08.1823049Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-12-04T09:25:08.1824176Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-12-04T09:25:08.2001388Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-12-04T09:25:08.2058401Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-12-04T09:25:08.2059082Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-12-04T09:25:08.2059678Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-12-04T09:25:08.2060328Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:25:08.2068314Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:25:08.2069165Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-12-04T09:25:08.2069985Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:25:08.2070882Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:25:08.2071607Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:25:08.2074076Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:25:08.2075522Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:25:08.2076707Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:25:08.2077638Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:25:08.2078499Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:25:08.2081354Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:25:08.2082684Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:25:08.2084054Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:25:08.2086347Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:25:08.2088264Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:25:08.2089278Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:25:08.2090134Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:25:08.2150628Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:25:08.2211640Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:25:08.2212892Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:25:08.2278472Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:25:08.2279684Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:25:08.2280960Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:25:08.2282340Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:25:08.2283662Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:25:08.2284949Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:25:08.2286178Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:25:08.2287534Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:25:08.2288693Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:25:08.2289880Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:25:08.2291059Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:25:08.2292218Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:25:08.2293325Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:25:08.2294690Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:25:08.2297854Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:25:08.2372934Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:25:08.2373941Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:25:08.2449791Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:25:08.2450774Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:25:08.2451503Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:25:08.2452313Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-12-04T09:25:08.2452978Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-12-04T09:25:08.2453948Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-12-04T09:25:08.2455020Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-12-04T09:25:08.2456095Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-12-04T09:25:08.2457057Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-12-04T09:25:08.2458066Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-12-04T09:25:08.2459056Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-12-04T09:25:08.2460087Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-12-04T09:25:08.2461115Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-12-04T09:25:08.2462366Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-12-04T09:25:08.2485169Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-12-04T09:25:08.2549239Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-12-04T09:25:08.2550354Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:25:08.2551372Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:25:08.2552267Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-12-04T09:25:08.2553289Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-12-04T09:25:08.2555461Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-12-04T09:25:08.2556299Z inflating: build/custom_test_artifacts/jit-hook-build/detect_cuda_version.cc 2025-12-04T09:25:08.2559153Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-12-04T09:25:08.2560081Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-12-04T09:25:08.2561171Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-12-04T09:25:08.2600953Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-12-04T09:25:08.2601637Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-12-04T09:25:08.2602312Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-12-04T09:25:08.2603017Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:25:08.2610889Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:25:08.2611818Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-12-04T09:25:08.2612725Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:25:08.2613825Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:25:08.2614532Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:25:08.2615697Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:25:08.2617461Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:25:08.2618562Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:25:08.2619562Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:25:08.2620486Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:25:08.2622994Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:25:08.2624266Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:25:08.2625615Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:25:08.2627490Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:25:08.2630435Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:25:08.2631545Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:25:08.2632280Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:25:08.2692328Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:25:08.2753066Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:25:08.2754501Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:25:08.2819832Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:25:08.2821175Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:25:08.2822514Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:25:08.2823953Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:25:08.2825364Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:25:08.2826737Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:25:08.2827979Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:25:08.2829305Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:25:08.2830647Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:25:08.2831944Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:25:08.2833184Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:25:08.2834337Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:25:08.2839884Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:25:08.2840799Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:25:08.2841724Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:25:08.2913815Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:25:08.2915780Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:25:08.2989393Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:25:08.2990471Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:25:08.2991356Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:25:08.2992249Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-12-04T09:25:08.2993197Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-12-04T09:25:08.2994265Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-12-04T09:25:08.2995289Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-12-04T09:25:08.2996228Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-12-04T09:25:08.2997317Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-12-04T09:25:08.2998284Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-12-04T09:25:08.2999263Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-12-04T09:25:08.3000198Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-12-04T09:25:08.3001042Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-12-04T09:25:08.3001874Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-12-04T09:25:08.3004867Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-12-04T09:25:08.3127241Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-12-04T09:25:08.3128285Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-12-04T09:25:08.3129229Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-12-04T09:25:08.3130271Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-12-04T09:25:08.3131278Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-12-04T09:25:08.3132230Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-12-04T09:25:08.3133211Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-12-04T09:25:08.3134194Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-12-04T09:25:08.3135156Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-12-04T09:25:08.3136587Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-12-04T09:25:08.3137453Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-12-04T09:25:08.3158119Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-12-04T09:25:08.3214343Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-12-04T09:25:08.3215385Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:25:08.3216339Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:25:08.3217248Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-12-04T09:25:08.3218120Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-12-04T09:25:08.3220097Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-12-04T09:25:08.3220890Z inflating: build/custom_test_artifacts/custom-backend-build/detect_cuda_version.cc 2025-12-04T09:25:08.3223829Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-12-04T09:25:08.3224798Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-12-04T09:25:08.3225786Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-12-04T09:25:08.3329569Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-12-04T09:25:08.3370199Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-12-04T09:25:08.3370816Z creating: build/lib/ 2025-12-04T09:25:08.3454781Z inflating: build/lib/libprotobuf-lite.a 2025-12-04T09:25:08.3902223Z inflating: build/lib/libprotobuf.a 2025-12-04T09:25:08.4400910Z inflating: build/lib/libprotoc.a 2025-12-04T09:25:08.4410352Z inflating: build/lib/libpthreadpool.a 2025-12-04T09:25:08.4418827Z inflating: build/lib/libcpuinfo.a 2025-12-04T09:25:08.4427115Z inflating: build/lib/libcpuinfo_internals.a 2025-12-04T09:25:08.4428014Z inflating: build/lib/libclog.a 2025-12-04T09:25:08.4448120Z inflating: build/lib/libpytorch_qnnpack.a 2025-12-04T09:25:08.4450859Z inflating: build/lib/libnnpack_reference_layers.a 2025-12-04T09:25:08.4469366Z inflating: build/lib/libnnpack.a 2025-12-04T09:25:08.4658209Z inflating: build/lib/libmicrokernels-prod.a 2025-12-04T09:25:08.5541945Z inflating: build/lib/libmicrokernels-all.a 2025-12-04T09:25:08.5612516Z inflating: build/lib/libgtest.a 2025-12-04T09:25:08.5629915Z inflating: build/lib/libgmock.a 2025-12-04T09:25:08.5630905Z inflating: build/lib/libgtest_main.a 2025-12-04T09:25:08.5631792Z inflating: build/lib/libgmock_main.a 2025-12-04T09:25:08.5723381Z inflating: build/lib/libXNNPACK.a 2025-12-04T09:25:08.5799644Z inflating: build/lib/libbenchmark.a 2025-12-04T09:25:08.5800476Z inflating: build/lib/libbenchmark_main.a 2025-12-04T09:25:08.5801754Z inflating: build/lib/libjitprofiling.a 2025-12-04T09:25:08.5868282Z inflating: build/lib/libasmjit.a 2025-12-04T09:25:08.5876514Z inflating: build/lib/libittnotify.a 2025-12-04T09:25:08.7077923Z inflating: build/lib/libfbgemm.a 2025-12-04T09:25:08.7108625Z inflating: build/lib/libtensorpipe_uv.a 2025-12-04T09:25:08.7659916Z inflating: build/lib/libtensorpipe.a 2025-12-04T09:25:08.7907318Z inflating: build/lib/libtensorpipe_cuda.a 2025-12-04T09:25:08.8042903Z inflating: build/lib/libgloo.a 2025-12-04T09:25:08.8090409Z inflating: build/lib/libonnx_proto.a 2025-12-04T09:25:08.8537384Z inflating: build/lib/libgloo_cuda.a 2025-12-04T09:25:08.9260289Z inflating: build/lib/libonnx.a 2025-12-04T09:25:09.9583186Z inflating: build/lib/libdnnl.a 2025-12-04T09:25:09.9603038Z inflating: build/lib/libfmt.a 2025-12-04T09:25:10.0085136Z inflating: build/lib/libkineto.a 2025-12-04T09:25:10.0203670Z inflating: build/lib/libc10.so 2025-12-04T09:25:10.0253582Z inflating: build/lib/libc10_cuda.so 2025-12-04T09:25:10.0256291Z inflating: build/lib/libcaffe2_nvrtc.so 2025-12-04T09:25:10.0257920Z inflating: build/lib/libtorch_global_deps.so 2025-12-04T09:25:13.1587071Z inflating: build/lib/libtorch_cpu.so 2025-12-04T09:25:13.2363232Z inflating: build/lib/libtorch_nvshmem.so 2025-12-04T09:25:15.2166372Z inflating: build/lib/libtorch_cuda.so 2025-12-04T09:25:15.2168711Z inflating: build/lib/libtorch.so 2025-12-04T09:25:15.2220259Z inflating: build/lib/libtorch_cuda_linalg.so 2025-12-04T09:25:15.2292846Z inflating: build/lib/libtorchbind_test.so 2025-12-04T09:25:15.2311880Z inflating: build/lib/libjitbackend_test.so 2025-12-04T09:25:15.2335998Z inflating: build/lib/libbackend_with_compiler.so 2025-12-04T09:25:15.2363047Z inflating: build/lib/libaoti_custom_ops.so 2025-12-04T09:25:15.2366133Z inflating: build/lib/libc10d_cuda_test.so 2025-12-04T09:25:15.2370580Z inflating: build/lib/libshm.so 2025-12-04T09:25:15.4779232Z inflating: build/lib/libtorch_python.so 2025-12-04T09:25:15.4815690Z inflating: build/lib/libnnapi_backend.so 2025-12-04T09:25:15.4816073Z creating: build/bin/ 2025-12-04T09:25:15.5277007Z inflating: build/bin/protoc-3.13.0.0 2025-12-04T09:25:15.5737617Z inflating: build/bin/protoc 2025-12-04T09:25:15.5797049Z inflating: build/bin/c10_AllocatorConfig_test 2025-12-04T09:25:15.5854231Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-12-04T09:25:15.5912322Z inflating: build/bin/c10_DeviceGuard_test 2025-12-04T09:25:15.5969720Z inflating: build/bin/c10_Device_test 2025-12-04T09:25:15.6036152Z inflating: build/bin/c10_DispatchKeySet_test 2025-12-04T09:25:15.6091240Z inflating: build/bin/c10_StreamGuard_test 2025-12-04T09:25:15.6151402Z inflating: build/bin/c10_Scalar_test 2025-12-04T09:25:15.6213796Z inflating: build/bin/c10_SizesAndStrides_test 2025-12-04T09:25:15.6274539Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-12-04T09:25:15.6337446Z inflating: build/bin/c10_SymInt_test 2025-12-04T09:25:15.6399342Z inflating: build/bin/c10_InlineStreamGuard_test 2025-12-04T09:25:15.6455142Z inflating: build/bin/c10_ArrayRef_test 2025-12-04T09:25:15.6531967Z inflating: build/bin/c10_cow_test 2025-12-04T09:25:15.6587097Z inflating: build/bin/c10_ConstexprCrc_test 2025-12-04T09:25:15.6643036Z inflating: build/bin/c10_DeadlockDetection_test 2025-12-04T09:25:15.6702070Z inflating: build/bin/c10_Bitset_test 2025-12-04T09:25:15.6765778Z inflating: build/bin/c10_Enumerate_test 2025-12-04T09:25:15.6824587Z inflating: build/bin/c10_IntrusiveList_test 2025-12-04T09:25:15.6881563Z inflating: build/bin/c10_Half_test 2025-12-04T09:25:15.6943833Z inflating: build/bin/c10_LeftRight_test 2025-12-04T09:25:15.7003572Z inflating: build/bin/c10_NetworkFlow_test 2025-12-04T09:25:15.7058944Z inflating: build/bin/c10_Semaphore_test 2025-12-04T09:25:15.7115233Z inflating: build/bin/c10_Synchronized_test 2025-12-04T09:25:15.7176699Z inflating: build/bin/c10_ThreadLocal_test 2025-12-04T09:25:15.7234880Z inflating: build/bin/c10_TypeIndex_test 2025-12-04T09:25:15.7292714Z inflating: build/bin/c10_accumulate_test 2025-12-04T09:25:15.7355694Z inflating: build/bin/c10_bfloat16_test 2025-12-04T09:25:15.7411868Z inflating: build/bin/c10_bit_cast_test 2025-12-04T09:25:15.7473551Z inflating: build/bin/c10_complex_test 2025-12-04T09:25:15.7536895Z inflating: build/bin/c10_complex_math_test 2025-12-04T09:25:15.7592726Z inflating: build/bin/c10_error_test 2025-12-04T09:25:15.7651519Z inflating: build/bin/c10_exception_test 2025-12-04T09:25:15.7707743Z inflating: build/bin/c10_flags_test 2025-12-04T09:25:15.7763999Z inflating: build/bin/c10_generic_math_test 2025-12-04T09:25:15.7821392Z inflating: build/bin/c10_irange_test 2025-12-04T09:25:15.7881315Z inflating: build/bin/c10_lazy_test 2025-12-04T09:25:15.8050844Z inflating: build/bin/c10_intrusive_ptr_test 2025-12-04T09:25:15.8114615Z inflating: build/bin/c10_logging_test 2025-12-04T09:25:15.8171000Z inflating: build/bin/c10_nofatal_test 2025-12-04T09:25:15.8253253Z inflating: build/bin/c10_optional_test 2025-12-04T09:25:15.8312764Z inflating: build/bin/c10_registry_test 2025-12-04T09:25:15.8381204Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-12-04T09:25:15.8546439Z inflating: build/bin/c10_small_vector_test 2025-12-04T09:25:15.8620984Z inflating: build/bin/c10_string_util_test 2025-12-04T09:25:15.8667237Z inflating: build/bin/c10_ssize_test 2025-12-04T09:25:15.8723168Z inflating: build/bin/c10_tempfile_test 2025-12-04T09:25:15.8778326Z inflating: build/bin/c10_string_view_test 2025-12-04T09:25:15.8841317Z inflating: build/bin/c10_typeid_test 2025-12-04T09:25:15.8891561Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-12-04T09:25:15.8950254Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_thread_and_block_and_device 2025-12-04T09:25:15.9010559Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_stream 2025-12-04T09:25:15.9069171Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_from_2_processes 2025-12-04T09:25:15.9128002Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_1_var_test 2025-12-04T09:25:15.9183979Z inflating: build/bin/c10_cuda_CUDATest 2025-12-04T09:25:15.9243751Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_multiple_blocks 2025-12-04T09:25:15.9302714Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_blocks_and_threads 2025-12-04T09:25:15.9361718Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_same_block 2025-12-04T09:25:15.9980110Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-12-04T09:25:16.0620110Z inflating: build/bin/vec_test_all_types_AVX512 2025-12-04T09:25:16.1264648Z inflating: build/bin/vec_test_all_types_AVX2 2025-12-04T09:25:16.1371170Z inflating: build/bin/test_aoti_abi_check 2025-12-04T09:25:16.1426320Z inflating: build/bin/test_vec_half_DEFAULT 2025-12-04T09:25:16.1482427Z inflating: build/bin/test_vec_half_AVX512 2025-12-04T09:25:16.1538423Z inflating: build/bin/test_vec_half_AVX2 2025-12-04T09:25:16.1619325Z inflating: build/bin/Dict_test 2025-12-04T09:25:16.1677782Z inflating: build/bin/Dimname_test 2025-12-04T09:25:16.1749432Z inflating: build/bin/MaybeOwned_test 2025-12-04T09:25:16.1812566Z inflating: build/bin/NamedTensor_test 2025-12-04T09:25:16.1878129Z inflating: build/bin/apply_utils_test 2025-12-04T09:25:16.1943286Z inflating: build/bin/atest 2025-12-04T09:25:16.2013788Z inflating: build/bin/basic 2025-12-04T09:25:16.2074999Z inflating: build/bin/broadcast_test 2025-12-04T09:25:16.2131407Z inflating: build/bin/cpu_allocator_test 2025-12-04T09:25:16.2196172Z inflating: build/bin/cpu_generator_test 2025-12-04T09:25:16.2254774Z inflating: build/bin/cpu_profiling_allocator_test 2025-12-04T09:25:16.2355148Z inflating: build/bin/cpu_rng_test 2025-12-04T09:25:16.2412710Z inflating: build/bin/dlconvertor_test 2025-12-04T09:25:16.2476789Z inflating: build/bin/extension_backend_test 2025-12-04T09:25:16.2538346Z inflating: build/bin/half_test 2025-12-04T09:25:16.2643875Z inflating: build/bin/ivalue_test 2025-12-04T09:25:16.2699484Z inflating: build/bin/lazy_tensor_test 2025-12-04T09:25:16.2758600Z inflating: build/bin/math_kernel_test 2025-12-04T09:25:16.2818206Z inflating: build/bin/memory_format_test 2025-12-04T09:25:16.2878333Z inflating: build/bin/memory_overlapping_test 2025-12-04T09:25:16.2937754Z inflating: build/bin/mobile_memory_cleanup 2025-12-04T09:25:16.2999966Z inflating: build/bin/native_test 2025-12-04T09:25:16.3057670Z inflating: build/bin/operator_name_test 2025-12-04T09:25:16.3114558Z inflating: build/bin/operators_test 2025-12-04T09:25:16.3172670Z inflating: build/bin/packedtensoraccessor_test 2025-12-04T09:25:16.3247762Z inflating: build/bin/pow_test 2025-12-04T09:25:16.3310223Z inflating: build/bin/quantized_test 2025-12-04T09:25:16.3366465Z inflating: build/bin/reduce_ops_test 2025-12-04T09:25:16.3423288Z inflating: build/bin/reportMemoryUsage_test 2025-12-04T09:25:16.3485530Z inflating: build/bin/scalar_tensor_test 2025-12-04T09:25:16.3549475Z inflating: build/bin/scalar_test 2025-12-04T09:25:16.3607029Z inflating: build/bin/StorageUtils_test 2025-12-04T09:25:16.3665164Z inflating: build/bin/stride_properties_test 2025-12-04T09:25:16.3751891Z inflating: build/bin/tensor_iterator_test 2025-12-04T09:25:16.3812224Z inflating: build/bin/test_parallel 2025-12-04T09:25:16.3869762Z inflating: build/bin/thread_init_test 2025-12-04T09:25:16.3931128Z inflating: build/bin/type_ptr_test 2025-12-04T09:25:16.3997383Z inflating: build/bin/type_test 2025-12-04T09:25:16.4056166Z inflating: build/bin/undefined_tensor_test 2025-12-04T09:25:16.4112978Z inflating: build/bin/verify_api_visibility 2025-12-04T09:25:16.4190785Z inflating: build/bin/legacy_vmap_test 2025-12-04T09:25:16.4248221Z inflating: build/bin/weakref_test 2025-12-04T09:25:16.4305543Z inflating: build/bin/wrapdim_test 2025-12-04T09:25:16.4363955Z inflating: build/bin/xla_tensor_test 2025-12-04T09:25:16.4430067Z inflating: build/bin/IListRef_test 2025-12-04T09:25:16.4545011Z inflating: build/bin/List_test 2025-12-04T09:25:16.4618294Z inflating: build/bin/KernelFunction_test 2025-12-04T09:25:16.4747557Z inflating: build/bin/kernel_function_legacy_test 2025-12-04T09:25:16.4851678Z inflating: build/bin/kernel_function_test 2025-12-04T09:25:16.4987258Z inflating: build/bin/kernel_lambda_legacy_test 2025-12-04T09:25:16.5098280Z inflating: build/bin/kernel_lambda_test 2025-12-04T09:25:16.5165452Z inflating: build/bin/kernel_stackbased_test 2025-12-04T09:25:16.5268346Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-12-04T09:25:16.5325115Z inflating: build/bin/CppSignature_test 2025-12-04T09:25:16.5386620Z inflating: build/bin/backend_fallback_test 2025-12-04T09:25:16.5441930Z inflating: build/bin/op_allowlist_test 2025-12-04T09:25:16.5772810Z inflating: build/bin/op_registration_test 2025-12-04T09:25:16.5847877Z inflating: build/bin/inline_container_test 2025-12-04T09:25:16.5907135Z inflating: build/bin/cuda_allocator_test 2025-12-04T09:25:16.5967188Z inflating: build/bin/cuda_apply_test 2025-12-04T09:25:16.6034129Z inflating: build/bin/cuda_atomic_ops_test 2025-12-04T09:25:16.6096540Z inflating: build/bin/cuda_caching_host_allocator_test 2025-12-04T09:25:16.6173811Z inflating: build/bin/cuda_complex_math_test 2025-12-04T09:25:16.6240587Z inflating: build/bin/cuda_complex_test 2025-12-04T09:25:16.6309169Z inflating: build/bin/cuda_cub_test 2025-12-04T09:25:16.6368996Z inflating: build/bin/cuda_cublas_handle_pool_test 2025-12-04T09:25:16.6424888Z inflating: build/bin/cuda_device_test 2025-12-04T09:25:16.6496396Z inflating: build/bin/cuda_distributions_test 2025-12-04T09:25:16.6556858Z inflating: build/bin/cuda_event_test 2025-12-04T09:25:16.6614216Z inflating: build/bin/cuda_dlconvertor_test 2025-12-04T09:25:16.6669753Z inflating: build/bin/cuda_exchange_device_test 2025-12-04T09:25:16.6728888Z inflating: build/bin/cuda_reportMemoryUsage_test 2025-12-04T09:25:16.6784750Z inflating: build/bin/cuda_allocatorTraceTracker_test 2025-12-04T09:25:16.6842168Z inflating: build/bin/cuda_integer_divider_test 2025-12-04T09:25:16.6910163Z inflating: build/bin/cuda_stream_test 2025-12-04T09:25:16.6966544Z inflating: build/bin/cuda_cudnn_test 2025-12-04T09:25:16.7022260Z inflating: build/bin/cuda_half_test 2025-12-04T09:25:16.7085425Z inflating: build/bin/cuda_generator_test 2025-12-04T09:25:16.7141443Z inflating: build/bin/cuda_optional_test 2025-12-04T09:25:16.7199376Z inflating: build/bin/cuda_packedtensoraccessor_test 2025-12-04T09:25:16.7258110Z inflating: build/bin/cuda_vectorized_test 2025-12-04T09:25:16.8396680Z inflating: build/bin/test_jit 2025-12-04T09:25:16.8455349Z inflating: build/bin/BackoffTest 2025-12-04T09:25:16.8514550Z inflating: build/bin/FileStoreTest 2025-12-04T09:25:16.8883032Z inflating: build/bin/test_lazy 2025-12-04T09:25:16.8946963Z inflating: build/bin/TCPStoreTest 2025-12-04T09:25:16.9007309Z inflating: build/bin/HashStoreTest 2025-12-04T09:25:16.9020944Z inflating: build/bin/ProcessGroupMPITest 2025-12-04T09:25:16.9024132Z inflating: build/bin/example_allreduce 2025-12-04T09:25:16.9098340Z inflating: build/bin/ProcessGroupGlooTest 2025-12-04T09:25:16.9161919Z inflating: build/bin/ProcessGroupGlooAsyncTest 2025-12-04T09:25:16.9232331Z inflating: build/bin/ProcessGroupNCCLTest 2025-12-04T09:25:16.9300878Z inflating: build/bin/ProcessGroupNCCLErrorsTest 2025-12-04T09:25:16.9362487Z inflating: build/bin/test_dist_autograd 2025-12-04T09:25:16.9438091Z inflating: build/bin/test_cpp_rpc 2025-12-04T09:25:16.9440779Z inflating: build/bin/parallel_benchmark 2025-12-04T09:25:17.0661801Z inflating: build/bin/test_api 2025-12-04T09:25:17.0665674Z inflating: build/bin/torch_shm_manager 2025-12-04T09:25:17.0666075Z creating: .additional_ci_files/ 2025-12-04T09:25:17.0731581Z inflating: .additional_ci_files/test-times.json 2025-12-04T09:25:17.0970198Z inflating: .additional_ci_files/test-class-times.json 2025-12-04T09:25:17.1025494Z ##[group]Run rm artifacts.zip 2025-12-04T09:25:17.1025788Z rm artifacts.zip 2025-12-04T09:25:17.1035603Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:17.1035983Z env: 2025-12-04T09:25:17.1036383Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:17.1036659Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:17.1036968Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:17.1037347Z ##[endgroup] 2025-12-04T09:25:17.3923931Z ##[group]Run df -H 2025-12-04T09:25:17.3924171Z df -H 2025-12-04T09:25:17.3933383Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:17.3933742Z env: 2025-12-04T09:25:17.3933945Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:17.3934209Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:17.3934517Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:17.3934850Z ##[endgroup] 2025-12-04T09:25:17.3991763Z Filesystem Size Used Avail Use% Mounted on 2025-12-04T09:25:17.3992138Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-12-04T09:25:17.3992466Z tmpfs 34G 0 34G 0% /dev/shm 2025-12-04T09:25:17.3992788Z tmpfs 14G 562k 14G 1% /run 2025-12-04T09:25:17.3993106Z /dev/nvme0n1p1 161G 54G 108G 34% / 2025-12-04T09:25:17.3993434Z tmpfs 34G 17k 34G 1% /tmp 2025-12-04T09:25:17.3993763Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-12-04T09:25:17.3994127Z tmpfs 6.7G 0 6.7G 0% /run/user/0 2025-12-04T09:25:17.4028154Z Prepare all required actions 2025-12-04T09:25:17.4029084Z Getting action download info 2025-12-04T09:25:17.6031050Z ##[group]Run ./.github/actions/download-td-artifacts 2025-12-04T09:25:17.6031392Z with: 2025-12-04T09:25:17.6031597Z env: 2025-12-04T09:25:17.6031812Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:17.6032083Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:17.6032406Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:17.6032763Z ##[endgroup] 2025-12-04T09:25:17.6123549Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:25:17.6123887Z with: 2025-12-04T09:25:17.6124093Z name: td_results 2025-12-04T09:25:17.6124335Z s3-bucket: gha-artifacts 2025-12-04T09:25:17.6124601Z region: us-east-1 2025-12-04T09:25:17.6124823Z env: 2025-12-04T09:25:17.6125036Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:17.6125301Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:17.6125806Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:17.6126203Z ##[endgroup] 2025-12-04T09:25:18.2760028Z (node:59474) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:25:18.2760498Z 2025-12-04T09:25:18.2760689Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:25:18.2761309Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:25:18.2762310Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:25:18.3807591Z Found 1 objects with prefix pytorch/pytorch/19922826259/td_results/ 2025-12-04T09:25:18.3808184Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/td_results.json 2025-12-04T09:25:18.4825712Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/td_results.json 2025-12-04T09:25:18.4830716Z Artifact download has finished successfully 2025-12-04T09:25:18.5168787Z ##[group]Run mkdir -p .additional_ci_files 2025-12-04T09:25:18.5169177Z mkdir -p .additional_ci_files 2025-12-04T09:25:18.5169593Z mv td_results.json .additional_ci_files/td_results.json || true 2025-12-04T09:25:18.5179573Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:18.5179925Z env: 2025-12-04T09:25:18.5180130Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:18.5180387Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:18.5180691Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:18.5181026Z ##[endgroup] 2025-12-04T09:25:18.5294717Z ##[group]Run .github/scripts/parse_ref.py 2025-12-04T09:25:18.5295094Z .github/scripts/parse_ref.py 2025-12-04T09:25:18.5303602Z shell: /usr/bin/bash -e {0} 2025-12-04T09:25:18.5303866Z env: 2025-12-04T09:25:18.5304090Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:18.5304354Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:18.5304656Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:18.5305022Z ##[endgroup] 2025-12-04T09:25:18.5554547Z Setting output branch=main 2025-12-04T09:25:18.5676631Z Prepare all required actions 2025-12-04T09:25:18.5676992Z Getting action download info 2025-12-04T09:25:18.7294400Z ##[group]Run ./.github/actions/filter-test-configs 2025-12-04T09:25:18.7294720Z with: 2025-12-04T09:25:18.7295094Z github-token: *** 2025-12-04T09:25:18.7303159Z test-matrix: {"include": [{"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}]} 2025-12-04T09:25:18.7312137Z job-name: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:25:18.7312828Z env: 2025-12-04T09:25:18.7313034Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:18.7313301Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:18.7313614Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:18.7313955Z ##[endgroup] 2025-12-04T09:25:18.7349700Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:25:18.7350042Z with: 2025-12-04T09:25:18.7350241Z shell: bash 2025-12-04T09:25:18.7350454Z timeout_minutes: 10 2025-12-04T09:25:18.7350694Z max_attempts: 5 2025-12-04T09:25:18.7350915Z retry_wait_seconds: 30 2025-12-04T09:25:18.7351690Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:25:18.7352668Z polling_interval_seconds: 1 2025-12-04T09:25:18.7352939Z warning_on_retry: true 2025-12-04T09:25:18.7353193Z continue_on_error: false 2025-12-04T09:25:18.7353442Z env: 2025-12-04T09:25:18.7353642Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:18.7353895Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:18.7354201Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:18.7354675Z GITHUB_TOKEN: *** 2025-12-04T09:25:18.7354896Z ##[endgroup] 2025-12-04T09:25:18.8388898Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:25:19.0746827Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:25:19.1962805Z Collecting requests==2.27.1 2025-12-04T09:25:19.2116887Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-12-04T09:25:19.4065623Z Collecting pyyaml==6.0.2 2025-12-04T09:25:19.4105402Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-12-04T09:25:19.4338526Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-12-04T09:25:19.8709846Z Collecting charset-normalizer~=2.0.0 2025-12-04T09:25:19.8751892Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-12-04T09:25:19.8797971Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-12-04T09:25:19.9308718Z Collecting certifi>=2017.4.17 2025-12-04T09:25:19.9342980Z Downloading certifi-2025.11.12-py3-none-any.whl (159 kB) 2025-12-04T09:25:20.0217823Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-12-04T09:25:20.1420727Z Successfully installed certifi-2025.11.12 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-12-04T09:25:20.8138570Z Command completed after 1 attempt(s). 2025-12-04T09:25:20.8205421Z ##[group]Run set -x 2025-12-04T09:25:20.8205705Z set -x 2025-12-04T09:25:20.8206216Z  2025-12-04T09:25:20.8206604Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:25:20.8207063Z # in runner workspace 2025-12-04T09:25:20.8207549Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-12-04T09:25:20.8216724Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:20.8217087Z env: 2025-12-04T09:25:20.8217298Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:20.8217563Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:20.8217879Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:20.8218224Z ##[endgroup] 2025-12-04T09:25:20.8250397Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-12-04T09:25:20.8432808Z Setting output branch=main 2025-12-04T09:25:20.8501552Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:25:20.8501957Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:25:20.8502312Z echo "Job name: ${JOB_NAME}" 2025-12-04T09:25:20.8502595Z  2025-12-04T09:25:20.8502961Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:25:20.8503419Z # in runner workspace 2025-12-04T09:25:20.8503843Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-12-04T09:25:20.8504294Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-12-04T09:25:20.8504623Z  --job-name "${JOB_NAME}" \ 2025-12-04T09:25:20.8512938Z  --test-matrix "{"include": [{"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}]}" \ 2025-12-04T09:25:20.8521621Z  --selected-test-configs "" \ 2025-12-04T09:25:20.8521955Z  --pr-number "${PR_NUMBER}" \ 2025-12-04T09:25:20.8522306Z  --tag "${TAG}" \ 2025-12-04T09:25:20.8522602Z  --event-name "${EVENT_NAME}" \ 2025-12-04T09:25:20.8522920Z  --schedule "${SCHEDULE}" \ 2025-12-04T09:25:20.8523222Z  --branch "${HEAD_BRANCH}" 2025-12-04T09:25:20.8532301Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:20.8532660Z env: 2025-12-04T09:25:20.8532862Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:20.8533120Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:20.8533430Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:20.8533975Z GITHUB_TOKEN: *** 2025-12-04T09:25:20.8534644Z JOB_NAME: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:25:20.8535614Z PR_NUMBER: 2025-12-04T09:25:20.8535825Z TAG: 2025-12-04T09:25:20.8536035Z EVENT_NAME: schedule 2025-12-04T09:25:20.8536269Z SCHEDULE: 29 8 * * * 2025-12-04T09:25:20.8536504Z HEAD_BRANCH: main 2025-12-04T09:25:20.8536731Z ##[endgroup] 2025-12-04T09:25:20.8567030Z Workflow: periodic 2025-12-04T09:25:20.8567946Z Job name: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:25:21.0334068Z Setting output keep-going=True 2025-12-04T09:25:21.0334439Z Setting output ci-verbose-test-logs=False 2025-12-04T09:25:21.0334789Z Setting output ci-test-showlocals=False 2025-12-04T09:25:21.0335120Z Setting output ci-no-test-timeout=False 2025-12-04T09:25:21.0335681Z Setting output ci-no-td=False 2025-12-04T09:25:21.0336222Z Setting output ci-td-distributed=False 2025-12-04T09:25:21.0336552Z Setting output is-unstable=False 2025-12-04T09:25:21.0336852Z Setting output reenabled-issues= 2025-12-04T09:25:21.0355154Z Setting output test-matrix={"include": [{"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}]} 2025-12-04T09:25:21.0373592Z Setting output is-test-matrix-empty=False 2025-12-04T09:25:21.0461359Z ##[group]Run echo "Filtered matrix:" 2025-12-04T09:25:21.0461728Z echo "Filtered matrix:" 2025-12-04T09:25:21.0479891Z echo "{"include": [{"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 7, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "mem_leak_check": "mem_leak_check", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 8, "num_shards": 8, "runner": "linux.g5.4xlarge.nvidia.gpu", "owners": ["module:slowgradcheck"], "rerun_disabled_tests": "rerun_disabled_tests"}]}" 2025-12-04T09:25:21.0498317Z  2025-12-04T09:25:21.0498515Z echo 2025-12-04T09:25:21.0498784Z echo "Is the current job unstable? False" 2025-12-04T09:25:21.0499100Z  2025-12-04T09:25:21.0499296Z echo 2025-12-04T09:25:21.0499543Z echo "Is keep-going label set? True" 2025-12-04T09:25:21.0499842Z  2025-12-04T09:25:21.0500035Z echo 2025-12-04T09:25:21.0500269Z echo "Reenabled issues? " 2025-12-04T09:25:21.0509379Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:21.0509736Z env: 2025-12-04T09:25:21.0509946Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:21.0510197Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:21.0510503Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:21.0510852Z ##[endgroup] 2025-12-04T09:25:21.0540160Z Filtered matrix: 2025-12-04T09:25:21.0562891Z {include: [{config: default, shard: 1, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check}, {config: default, shard: 1, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 1, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 1, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 2, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check}, {config: default, shard: 2, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 2, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 2, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 3, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check}, {config: default, shard: 3, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 3, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 3, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 4, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check}, {config: default, shard: 4, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 4, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 4, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 5, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check}, {config: default, shard: 5, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 5, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 5, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 6, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check}, {config: default, shard: 6, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 6, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 6, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 7, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check}, {config: default, shard: 7, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 7, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 7, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 8, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check}, {config: default, shard: 8, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], mem_leak_check: mem_leak_check, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 8, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests, mem_leak_check: mem_leak_check}, {config: default, shard: 8, num_shards: 8, runner: linux.g5.4xlarge.nvidia.gpu, owners: [module:slowgradcheck], rerun_disabled_tests: rerun_disabled_tests}]} 2025-12-04T09:25:21.0581014Z 2025-12-04T09:25:21.0581126Z Is the current job unstable? False 2025-12-04T09:25:21.0581336Z 2025-12-04T09:25:21.0581452Z Is keep-going label set? True 2025-12-04T09:25:21.0581665Z 2025-12-04T09:25:21.0581778Z Reenabled issues? 2025-12-04T09:25:21.0615892Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:25:21.0616420Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:25:21.0625494Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:21.0625853Z env: 2025-12-04T09:25:21.0626073Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:21.0626340Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:21.0626648Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:21.0626997Z JOB_TIMEOUT: 300 2025-12-04T09:25:21.0627228Z ##[endgroup] 2025-12-04T09:25:21.0684017Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:25:21.0684527Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:25:21.0684967Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:25:21.0693489Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:25:21.0693873Z env: 2025-12-04T09:25:21.0694089Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:21.0694357Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:21.0694663Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:21.0695018Z ##[endgroup] 2025-12-04T09:25:21.0818780Z ##[group]Run set -x 2025-12-04T09:25:21.0819114Z set -x 2025-12-04T09:25:21.0819337Z  2025-12-04T09:25:21.0819578Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-12-04T09:25:21.0819964Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-12-04T09:25:21.0820355Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-12-04T09:25:21.0820714Z  TEST_COMMAND=.ci/onnx/test.sh 2025-12-04T09:25:21.0821001Z else 2025-12-04T09:25:21.0821247Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:25:21.0821543Z fi 2025-12-04T09:25:21.0821736Z  2025-12-04T09:25:21.0821993Z # Leaving 1GB for the runner and other things 2025-12-04T09:25:21.0822564Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-12-04T09:25:21.0823378Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-12-04T09:25:21.0824047Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-12-04T09:25:21.0824551Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-12-04T09:25:21.0824946Z  2025-12-04T09:25:21.0825207Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:25:21.0825530Z  SHM_OPTS= 2025-12-04T09:25:21.0825777Z  JENKINS_USER= 2025-12-04T09:25:21.0826124Z  # ensure that docker container cleanly exits in 12 hours 2025-12-04T09:25:21.0826581Z  # if for some reason cleanup action doesn't stop container 2025-12-04T09:25:21.0826971Z  # when job is cancelled 2025-12-04T09:25:21.0827449Z  DOCKER_SHELL_CMD="sleep 12h" 2025-12-04T09:25:21.0827773Z  USED_IMAGE="${DOCKER_IMAGE_S390X}" 2025-12-04T09:25:21.0828072Z else 2025-12-04T09:25:21.0828320Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-12-04T09:25:21.0828668Z  JENKINS_USER="--user jenkins" 2025-12-04T09:25:21.0828968Z  DOCKER_SHELL_CMD= 2025-12-04T09:25:21.0829258Z  USED_IMAGE="${DOCKER_IMAGE}" 2025-12-04T09:25:21.0829545Z fi 2025-12-04T09:25:21.0829743Z  2025-12-04T09:25:21.0830079Z # detached container should get cleaned up by teardown_ec2_linux 2025-12-04T09:25:21.0830596Z # TODO: Stop building test binaries as part of the build phase 2025-12-04T09:25:21.0831188Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-12-04T09:25:21.0831707Z # shellcheck disable=SC2086,SC2090 2025-12-04T09:25:21.0832039Z container_name=$(docker run \ 2025-12-04T09:25:21.0832355Z  ${GPU_FLAG:-} \ 2025-12-04T09:25:21.0832649Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-12-04T09:25:21.0832999Z  -e BUILD_ENVIRONMENT \ 2025-12-04T09:25:21.0833294Z  -e PR_NUMBER \ 2025-12-04T09:25:21.0833575Z  -e GITHUB_ACTIONS \ 2025-12-04T09:25:21.0833861Z  -e GITHUB_REPOSITORY \ 2025-12-04T09:25:21.0834166Z  -e GITHUB_WORKFLOW \ 2025-12-04T09:25:21.0834450Z  -e GITHUB_JOB \ 2025-12-04T09:25:21.0834702Z  -e GITHUB_RUN_ID \ 2025-12-04T09:25:21.0834979Z  -e GITHUB_RUN_NUMBER \ 2025-12-04T09:25:21.0835473Z  -e GITHUB_RUN_ATTEMPT \ 2025-12-04T09:25:21.0835753Z  -e JOB_ID \ 2025-12-04T09:25:21.0835999Z  -e JOB_NAME \ 2025-12-04T09:25:21.0836259Z  -e BASE_SHA \ 2025-12-04T09:25:21.0836512Z  -e BRANCH \ 2025-12-04T09:25:21.0836757Z  -e SHA1 \ 2025-12-04T09:25:21.0837009Z  -e AWS_DEFAULT_REGION \ 2025-12-04T09:25:21.0837309Z  -e IN_WHEEL_TEST \ 2025-12-04T09:25:21.0837572Z  -e SHARD_NUMBER \ 2025-12-04T09:25:21.0837838Z  -e TEST_CONFIG \ 2025-12-04T09:25:21.0838123Z  -e NUM_TEST_SHARDS \ 2025-12-04T09:25:21.0838562Z  -e REENABLED_ISSUES \ 2025-12-04T09:25:21.0838869Z  -e CONTINUE_THROUGH_ERROR \ 2025-12-04T09:25:21.0839177Z  -e VERBOSE_TEST_LOGS \ 2025-12-04T09:25:21.0839459Z  -e TEST_SHOWLOCALS \ 2025-12-04T09:25:21.0839742Z  -e NO_TEST_TIMEOUT \ 2025-12-04T09:25:21.0840016Z  -e NO_TD \ 2025-12-04T09:25:21.0840266Z  -e TD_DISTRIBUTED \ 2025-12-04T09:25:21.0840542Z  -e PR_LABELS \ 2025-12-04T09:25:21.0840831Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-12-04T09:25:21.0841155Z  -e SCCACHE_BUCKET \ 2025-12-04T09:25:21.0841426Z  -e SCCACHE_REGION \ 2025-12-04T09:25:21.0841727Z  -e XLA_CUDA \ 2025-12-04T09:25:21.0842037Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-12-04T09:25:21.0842387Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-12-04T09:25:21.0842756Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-12-04T09:25:21.0843116Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-12-04T09:25:21.0843455Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-12-04T09:25:21.0843773Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-12-04T09:25:21.0844115Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-12-04T09:25:21.0844439Z  -e DASHBOARD_TAG \ 2025-12-04T09:25:21.0844724Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-12-04T09:25:21.0845089Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-12-04T09:25:21.0845510Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-12-04T09:25:21.0845911Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-12-04T09:25:21.0846290Z  --security-opt seccomp=unconfined \ 2025-12-04T09:25:21.0846636Z  --cap-add=SYS_PTRACE \ 2025-12-04T09:25:21.0847064Z  --ipc=host \ 2025-12-04T09:25:21.0847397Z  ${SHM_OPTS} \ 2025-12-04T09:25:21.0847659Z  --tty \ 2025-12-04T09:25:21.0847894Z  --detach \ 2025-12-04T09:25:21.0848163Z  --name="${container_name}" \ 2025-12-04T09:25:21.0848471Z  ${JENKINS_USER} \ 2025-12-04T09:25:21.0848814Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-12-04T09:25:21.0849224Z  -w /var/lib/jenkins/workspace \ 2025-12-04T09:25:21.0849532Z  "${USED_IMAGE}" \ 2025-12-04T09:25:21.0849811Z  ${DOCKER_SHELL_CMD} 2025-12-04T09:25:21.0850073Z ) 2025-12-04T09:25:21.0850402Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-12-04T09:25:21.0850801Z  2025-12-04T09:25:21.0851045Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:25:21.0851764Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-12-04T09:25:21.0852272Z fi 2025-12-04T09:25:21.0852496Z  2025-12-04T09:25:21.0852988Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-12-04T09:25:21.0862204Z shell: /usr/bin/bash -e {0} 2025-12-04T09:25:21.0862478Z env: 2025-12-04T09:25:21.0862697Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:25:21.0862964Z HAS_NVIDIA_GPU: true 2025-12-04T09:25:21.0863295Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:21.0863790Z BUILD_ENVIRONMENT: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck 2025-12-04T09:25:21.0864199Z PR_NUMBER: 2025-12-04T09:25:21.0864454Z GITHUB_REPOSITORY: pytorch/pytorch 2025-12-04T09:25:21.0864773Z GITHUB_WORKFLOW: periodic 2025-12-04T09:25:21.0865046Z GITHUB_JOB: test 2025-12-04T09:25:21.0865290Z GITHUB_RUN_ID: 19922826259 2025-12-04T09:25:21.0865575Z GITHUB_RUN_NUMBER: 19107 2025-12-04T09:25:21.0865858Z GITHUB_RUN_ATTEMPT: 1 2025-12-04T09:25:21.0866096Z JOB_ID: 57118183167 2025-12-04T09:25:21.0866897Z JOB_NAME: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:25:21.0867609Z BRANCH: main 2025-12-04T09:25:21.0867883Z SHA1: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:21.0868277Z BASE_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:21.0868619Z TEST_CONFIG: default 2025-12-04T09:25:21.0868867Z SHARD_NUMBER: 2 2025-12-04T09:25:21.0869090Z NUM_TEST_SHARDS: 8 2025-12-04T09:25:21.0869328Z EXTRA_FLAGS: 2025-12-04T09:25:21.0869565Z OP_BENCHMARK_TESTS: 2025-12-04T09:25:21.0869809Z REENABLED_ISSUES: 2025-12-04T09:25:21.0870072Z CONTINUE_THROUGH_ERROR: True 2025-12-04T09:25:21.0870359Z VERBOSE_TEST_LOGS: False 2025-12-04T09:25:21.0870614Z TEST_SHOWLOCALS: False 2025-12-04T09:25:21.0870880Z NO_TEST_TIMEOUT: False 2025-12-04T09:25:21.0871136Z NO_TD: False 2025-12-04T09:25:21.0871362Z TD_DISTRIBUTED: False 2025-12-04T09:25:21.0871676Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-12-04T09:25:21.0872036Z SCCACHE_REGION: us-east-1 2025-12-04T09:25:21.0872288Z SHM_SIZE: 2g 2025-12-04T09:25:21.0873041Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:25:21.0874375Z DOCKER_IMAGE_S390X: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:25:21.0875185Z XLA_CUDA: 2025-12-04T09:25:21.0875532Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:25:21.0875980Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-12-04T09:25:21.0876297Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 1 2025-12-04T09:25:21.0876599Z DASHBOARD_TAG: 2025-12-04T09:25:21.0877008Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-12-04T09:25:21.0877510Z HUGGING_FACE_HUB_TOKEN: *** 2025-12-04T09:25:21.0877918Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-12-04T09:25:21.0878380Z ARTIFACTS_FILE_SUFFIX: test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167 2025-12-04T09:25:21.0878837Z ##[endgroup] 2025-12-04T09:25:21.0909074Z + [[ default == \m\u\l\t\i\g\p\u ]] 2025-12-04T09:25:21.0909632Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *onnx* ]] 2025-12-04T09:25:21.0910190Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:25:21.0913487Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-12-04T09:25:21.0938799Z + TOTAL_AVAILABLE_MEMORY_IN_GB='61.094 ' 2025-12-04T09:25:21.0939272Z + TOTAL_MEMORY_WITH_SWAP=64 2025-12-04T09:25:21.0939864Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *\s\3\9\0\x* ]] 2025-12-04T09:25:21.0940485Z + SHM_OPTS=--shm-size=2g 2025-12-04T09:25:21.0940872Z + JENKINS_USER='--user jenkins' 2025-12-04T09:25:21.0941277Z + DOCKER_SHELL_CMD= 2025-12-04T09:25:21.0942311Z + USED_IMAGE=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:25:21.0950802Z +++ nproc --ignore=2 2025-12-04T09:25:21.0980989Z ++ docker run --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=14 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=61g --memory-swap=64g --env-file=/tmp/github_env_19922826259 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=2g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:25:29.5765362Z + container_name=45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:25:29.5766418Z + echo DOCKER_CONTAINER_ID=45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:25:29.5767499Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *\s\3\9\0\x* ]] 2025-12-04T09:25:29.5774810Z ++ echo dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:25:29.5778582Z + docker exec -t 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 sh -c 'python3 -m pip install dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-12-04T09:25:30.0440369Z Processing ./dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl (from torch==2.10.0a0+gitffd9b0f) 2025-12-04T09:25:30.3676055Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.18.0) 2025-12-04T09:25:30.3678007Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (4.12.2) 2025-12-04T09:25:30.3683063Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.13.3) 2025-12-04T09:25:30.3687753Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2.8.8) 2025-12-04T09:25:30.3691682Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.1.6) 2025-12-04T09:25:30.3696435Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2025.10.0) 2025-12-04T09:25:30.3709606Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.3.0) 2025-12-04T09:25:30.4097714Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from opt-einsum>=3.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.22.4) 2025-12-04T09:25:30.4116551Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.3.0) 2025-12-04T09:25:30.4174933Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.0.3) 2025-12-04T09:25:30.7924842Z Installing collected packages: torch 2025-12-04T09:25:42.1731261Z Successfully installed torch-2.10.0a0+gitffd9b0f 2025-12-04T09:25:42.2499928Z + export TERM=vt100 2025-12-04T09:25:42.2502576Z + TERM=vt100 2025-12-04T09:25:42.2502936Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:25:42.2513119Z + source .ci/pytorch/common.sh 2025-12-04T09:25:42.2516484Z +++ dirname .ci/pytorch/common.sh 2025-12-04T09:25:42.2525553Z ++ source .ci/pytorch/common_utils.sh 2025-12-04T09:25:42.2527764Z +++ declare -f -t trap_add 2025-12-04T09:25:42.2532195Z ++ set -ex -o pipefail 2025-12-04T09:25:42.2532683Z ++ [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *rocm* ]] 2025-12-04T09:25:42.2533089Z ++ BUILD_TEST_LIBTORCH=0 2025-12-04T09:25:42.2536802Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:25:42.2547366Z + source .ci/pytorch/common-build.sh 2025-12-04T09:25:42.2549806Z ++ [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *win-* ]] 2025-12-04T09:25:42.2556765Z ++++ dirname .ci/pytorch/common-build.sh 2025-12-04T09:25:42.2568288Z +++ cd .ci/pytorch 2025-12-04T09:25:42.2568728Z +++ pwd -P 2025-12-04T09:25:42.2571724Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-12-04T09:25:42.2572218Z ++ [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *-pch* ]] 2025-12-04T09:25:42.2572616Z ++ which sccache 2025-12-04T09:25:42.2642857Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-12-04T09:25:42.2643214Z ++ sccache --stop-server 2025-12-04T09:25:42.2674774Z ++ true 2025-12-04T09:25:42.2675185Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-12-04T09:25:42.2686998Z ++ trap_add sccache_epilogue EXIT 2025-12-04T09:25:42.2687373Z ++ trap_add_cmd=sccache_epilogue 2025-12-04T09:25:42.2687900Z ++ shift 2025-12-04T09:25:42.2688155Z ++ for trap_add_name in "$@" 2025-12-04T09:25:42.2694212Z ++++ trap -p EXIT 2025-12-04T09:25:42.2697718Z +++ eval 'extract_trap_cmd ' 2025-12-04T09:25:42.2697993Z ++++ extract_trap_cmd 2025-12-04T09:25:42.2698234Z ++++ printf '%s\n' '' 2025-12-04T09:25:42.2698491Z +++ printf '%s\n' sccache_epilogue 2025-12-04T09:25:42.2701497Z ++ trap -- ' 2025-12-04T09:25:42.2701874Z sccache_epilogue' EXIT 2025-12-04T09:25:42.2702143Z ++ [[ -n 1 ]] 2025-12-04T09:25:42.2702518Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-12-04T09:25:42.2703118Z Skipping sccache server initialization, setting environment variables 2025-12-04T09:25:42.2703553Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:25:42.2703836Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:25:42.2704177Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:25:42.2704604Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:25:42.2710485Z ++ export RUST_LOG=sccache::server=error 2025-12-04T09:25:42.2711062Z ++ RUST_LOG=sccache::server=error 2025-12-04T09:25:42.2711355Z ++ sccache --zero-stats 2025-12-04T09:25:42.6305993Z Statistics zeroed. 2025-12-04T09:25:42.6315045Z ++ which ccache 2025-12-04T09:25:42.6351010Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *rocm* ]] 2025-12-04T09:25:42.6351610Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *s390x* ]] 2025-12-04T09:25:42.6352021Z + [[ -d /var/lib/jenkins/workspace ]] 2025-12-04T09:25:42.6353932Z ++ stat -c %u /var/lib/jenkins/workspace 2025-12-04T09:25:42.6371645Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-12-04T09:25:42.6371974Z + trap_add cleanup_workspace EXIT 2025-12-04T09:25:42.6372283Z + trap_add_cmd=cleanup_workspace 2025-12-04T09:25:42.6372551Z + shift 2025-12-04T09:25:42.6372764Z + for trap_add_name in "$@" 2025-12-04T09:25:42.6378976Z +++ trap -p EXIT 2025-12-04T09:25:42.6382207Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-12-04T09:25:42.6382603Z sccache_epilogue'\'' EXIT' 2025-12-04T09:25:42.6382880Z +++ extract_trap_cmd trap -- ' 2025-12-04T09:25:42.6383150Z sccache_epilogue' EXIT 2025-12-04T09:25:42.6383468Z +++ printf '%s\n' ' 2025-12-04T09:25:42.6383731Z sccache_epilogue' 2025-12-04T09:25:42.6384008Z ++ printf '%s\n' cleanup_workspace 2025-12-04T09:25:42.6385093Z + trap -- ' 2025-12-04T09:25:43.6701494Z sccache_epilogue 2025-12-04T09:25:43.6701777Z cleanup_workspace' EXIT 2025-12-04T09:25:43.6702109Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-12-04T09:25:43.6702565Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-12-04T09:25:43.6724023Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *cuda* ]] 2025-12-04T09:25:43.6727537Z ++ python -c 'import os;import numba.cuda; print(os.path.dirname(numba.cuda.__file__))' 2025-12-04T09:25:44.1080261Z + NUMBA_CUDA_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:25:44.1080857Z + '[' -n /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda ']' 2025-12-04T09:25:44.1087446Z +++ realpath .ci/pytorch/test.sh 2025-12-04T09:25:44.1101247Z ++ dirname /var/lib/jenkins/workspace/.ci/pytorch/test.sh 2025-12-04T09:25:44.1254363Z + NUMBA_PATCH=/var/lib/jenkins/workspace/.ci/pytorch/numba-cuda-13.patch 2025-12-04T09:25:44.1255197Z + pushd /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:25:44.1255758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda ~/workspace 2025-12-04T09:25:44.1256191Z + patch -p4 2025-12-04T09:25:44.1270680Z patching file cudadrv/driver.py 2025-12-04T09:25:44.1270995Z Hunk #1 succeeded at 357 (offset -8 lines). 2025-12-04T09:25:44.1382771Z + popd 2025-12-04T09:25:44.1382987Z ~/workspace 2025-12-04T09:25:44.1383207Z + echo 'Environment variables:' 2025-12-04T09:25:44.1383494Z Environment variables: 2025-12-04T09:25:44.1383724Z + env 2025-12-04T09:25:44.1393846Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:25:44.1394483Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:25:44.1394907Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck 2025-12-04T09:25:44.1395618Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:25:44.1395905Z HOSTNAME=45a59d94bb2d 2025-12-04T09:25:44.1396465Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.1397059Z GITHUB_ACTION=__run_3 2025-12-04T09:25:44.1397334Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:25:44.1397622Z GITHUB_RUN_NUMBER=19107 2025-12-04T09:25:44.1397872Z TEST_CONFIG=default 2025-12-04T09:25:44.1398128Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:25:44.1398446Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:25:44.1398755Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:25:44.1399133Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:25:44.1399428Z GITHUB_TRIGGERING_ACTOR=huydhn 2025-12-04T09:25:44.1399699Z GITHUB_REF_TYPE=branch 2025-12-04T09:25:44.1400010Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:44.1400364Z XLA_CUDA= 2025-12-04T09:25:44.1400871Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:25:44.1401274Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:25:44.1401836Z *** 2025-12-04T09:25:44.1402057Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:25:44.1402334Z GITHUB_ACTIONS=true 2025-12-04T09:25:44.1402590Z NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:44.1413476Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:25:44.1413923Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:44.1414292Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:44.1414807Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/periodic.yml@refs/heads/main 2025-12-04T09:25:44.1415269Z UCC_HOME=/usr 2025-12-04T09:25:44.1415488Z VERBOSE_TEST_LOGS=False 2025-12-04T09:25:44.1415743Z GITHUB_REF=refs/heads/main 2025-12-04T09:25:44.1415992Z SHARD_NUMBER=2 2025-12-04T09:25:44.1416214Z GITHUB_REF_PROTECTED=true 2025-12-04T09:25:44.1416471Z HOME=/var/lib/jenkins 2025-12-04T09:25:44.1416745Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:25:44.1417078Z PYTORCH_TEST_RERUN_DISABLED_TESTS=1 2025-12-04T09:25:44.1417413Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-12-04T09:25:44.1417736Z USE_SYSTEM_NCCL=1 2025-12-04T09:25:44.1417964Z NUM_TEST_SHARDS=8 2025-12-04T09:25:44.1418179Z UCX_HOME=/usr 2025-12-04T09:25:44.1418728Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.1419758Z JOB_NAME=linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:25:44.1420750Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.1421521Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:25:44.1422002Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:25:44.1422252Z DASHBOARD_TAG= 2025-12-04T09:25:44.1422472Z GITHUB_RUN_ID=19922826259 2025-12-04T09:25:44.1422739Z INSTALLED_OPENBLAS= 2025-12-04T09:25:44.1423324Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.1424146Z GITHUB_ACTOR=huydhn 2025-12-04T09:25:44.1424402Z PR_NUMBER= 2025-12-04T09:25:44.1424609Z DESIRED_CUDA=12.8.1 2025-12-04T09:25:44.1424836Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:25:44.1425080Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:25:44.1425410Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:25:44.1425741Z TERM=vt100 2025-12-04T09:25:44.1425942Z INSTALLED_VISION=yes 2025-12-04T09:25:44.1426175Z BRANCH=main 2025-12-04T09:25:44.1426394Z SCCACHE_REGION=us-east-1 2025-12-04T09:25:44.1426658Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:25:44.1426935Z BUILD_AOT_INDUCTOR_TEST= 2025-12-04T09:25:44.1427191Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:25:44.1427688Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:25:44.1428288Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:25:44.1428632Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-12-04T09:25:44.1428955Z REENABLED_ISSUES= 2025-12-04T09:25:44.1429168Z DOCS= 2025-12-04T09:25:44.1429359Z SHLVL=1 2025-12-04T09:25:44.1429540Z MAX_JOBS=14 2025-12-04T09:25:44.1429755Z GITHUB_ACTOR_ID=475357 2025-12-04T09:25:44.1430085Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:44.1430451Z GITHUB_REF_NAME=main 2025-12-04T09:25:44.1430812Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:25:44.1431218Z GITHUB_JOB=test 2025-12-04T09:25:44.1431436Z NO_TEST_TIMEOUT=False 2025-12-04T09:25:44.1431666Z TD_DISTRIBUTED=False 2025-12-04T09:25:44.1431930Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:25:44.1432224Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:25:44.1432472Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:25:44.1432743Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:25:44.1433468Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:25:44.1434303Z GITHUB_BASE_REF= 2025-12-04T09:25:44.1434529Z INSTALLED_ACL= 2025-12-04T09:25:44.1434934Z ARTIFACTS_FILE_SUFFIX=test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167 2025-12-04T09:25:44.1435746Z CI=true 2025-12-04T09:25:44.1435978Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:25:44.1436295Z RUST_LOG=sccache::server=error 2025-12-04T09:25:44.1436559Z JOB_ID=57118183167 2025-12-04T09:25:44.1436772Z GITHUB_HEAD_REF= 2025-12-04T09:25:44.1436993Z GITHUB_ACTION_REF= 2025-12-04T09:25:44.1437274Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:25:44.1437609Z TEST_SHOWLOCALS=False 2025-12-04T09:25:44.1437859Z GITHUB_WORKFLOW=periodic 2025-12-04T09:25:44.1438125Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:25:44.1438709Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.1439318Z NO_TD=False 2025-12-04T09:25:44.1439544Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:25:44.1439839Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:25:44.1440136Z _=/usr/bin/env 2025-12-04T09:25:44.1440567Z OLDPWD=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:25:44.1441076Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-12-04T09:25:44.1539844Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch 2025-12-04T09:25:44.1540516Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:25:44.1541157Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib 2025-12-04T09:25:44.1541799Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/test 2025-12-04T09:25:44.1542313Z + BUILD_DIR=build 2025-12-04T09:25:44.1542559Z + BUILD_RENAMED_DIR=build_renamed 2025-12-04T09:25:44.1542887Z + BUILD_BIN_DIR=build/bin 2025-12-04T09:25:44.1543235Z + SHARD_NUMBER=2 2025-12-04T09:25:44.1543538Z + NUM_TEST_SHARDS=8 2025-12-04T09:25:44.1543799Z + export TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:25:44.1544097Z + TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:25:44.1544562Z + export VALGRIND=ON 2025-12-04T09:25:44.1544811Z + VALGRIND=ON 2025-12-04T09:25:44.1545139Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *clang9* ]] 2025-12-04T09:25:44.1545634Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *xpu* ]] 2025-12-04T09:25:44.1546005Z + detect_cuda_arch 2025-12-04T09:25:44.1546338Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *cuda* ]] 2025-12-04T09:25:44.1546724Z + command -v nvidia-smi 2025-12-04T09:25:44.1546964Z /usr/bin/nvidia-smi 2025-12-04T09:25:44.1551172Z ++ nvidia-smi --query-gpu=compute_cap --format=csv 2025-12-04T09:25:44.1552464Z ++ tail -n 1 2025-12-04T09:25:44.1810416Z + TORCH_CUDA_ARCH_LIST=8.6 2025-12-04T09:25:44.1810723Z + export TORCH_CUDA_ARCH_LIST 2025-12-04T09:25:44.1811116Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *s390x* ]] 2025-12-04T09:25:44.1811531Z + [[ 1 == \1 ]] 2025-12-04T09:25:44.1811753Z + ulimit -c 0 2025-12-04T09:25:44.1812077Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *bazel* ]] 2025-12-04T09:25:44.1815713Z ++ realpath build/custom_test_artifacts 2025-12-04T09:25:44.2007340Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-12-04T09:25:44.2007904Z + [[ -n '' ]] 2025-12-04T09:25:44.2008138Z + echo 'Environment variables' 2025-12-04T09:25:44.2008420Z Environment variables 2025-12-04T09:25:44.2008654Z + env 2025-12-04T09:25:44.2221152Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:25:44.2221757Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:25:44.2222290Z BUILD_ENVIRONMENT=linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck 2025-12-04T09:25:44.2223057Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:25:44.2223453Z HOSTNAME=45a59d94bb2d 2025-12-04T09:25:44.2224066Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.2225050Z GITHUB_ACTION=__run_3 2025-12-04T09:25:44.2225375Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:25:44.2225720Z GITHUB_RUN_NUMBER=19107 2025-12-04T09:25:44.2226052Z TEST_CONFIG=default 2025-12-04T09:25:44.2226371Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:25:44.2226719Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:25:44.2227030Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:25:44.2227423Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:25:44.2227720Z GITHUB_TRIGGERING_ACTOR=huydhn 2025-12-04T09:25:44.2227999Z GITHUB_REF_TYPE=branch 2025-12-04T09:25:44.2228247Z TORCH_CUDA_ARCH_LIST=8.6 2025-12-04T09:25:44.2228542Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:44.2228866Z XLA_CUDA= 2025-12-04T09:25:44.2229093Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:25:44.2229671Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:25:44.2229971Z *** 2025-12-04T09:25:44.2230190Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:25:44.2230465Z GITHUB_ACTIONS=true 2025-12-04T09:25:44.2230716Z NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:25:44.2231054Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:25:44.2231442Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:44.2231925Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:44.2232506Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/periodic.yml@refs/heads/main 2025-12-04T09:25:44.2232983Z UCC_HOME=/usr 2025-12-04T09:25:44.2233249Z TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:25:44.2233595Z VERBOSE_TEST_LOGS=False 2025-12-04T09:25:44.2233895Z GITHUB_REF=refs/heads/main 2025-12-04T09:25:44.2234201Z SHARD_NUMBER=2 2025-12-04T09:25:44.2234482Z GITHUB_REF_PROTECTED=true 2025-12-04T09:25:44.2234741Z HOME=/var/lib/jenkins 2025-12-04T09:25:44.2235066Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:25:44.2235714Z PYTORCH_TEST_RERUN_DISABLED_TESTS=1 2025-12-04T09:25:44.2236134Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-12-04T09:25:44.2236543Z USE_SYSTEM_NCCL=1 2025-12-04T09:25:44.2236760Z NUM_TEST_SHARDS=8 2025-12-04T09:25:44.2236986Z UCX_HOME=/usr 2025-12-04T09:25:44.2237736Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.2238780Z JOB_NAME=linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:25:44.2239789Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.2240559Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:25:44.2241043Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:25:44.2241301Z DASHBOARD_TAG= 2025-12-04T09:25:44.2241522Z GITHUB_RUN_ID=19922826259 2025-12-04T09:25:44.2241775Z INSTALLED_OPENBLAS= 2025-12-04T09:25:44.2242363Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.2242996Z GITHUB_ACTOR=huydhn 2025-12-04T09:25:44.2243222Z PR_NUMBER= 2025-12-04T09:25:44.2243429Z DESIRED_CUDA=12.8.1 2025-12-04T09:25:44.2243656Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:25:44.2243889Z VALGRIND=ON 2025-12-04T09:25:44.2244111Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:25:44.2244435Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:25:44.2244819Z TERM=vt100 2025-12-04T09:25:44.2245028Z INSTALLED_VISION=yes 2025-12-04T09:25:44.2245250Z BRANCH=main 2025-12-04T09:25:44.2245461Z SCCACHE_REGION=us-east-1 2025-12-04T09:25:44.2245725Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:25:44.2245989Z BUILD_AOT_INDUCTOR_TEST= 2025-12-04T09:25:44.2246243Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:25:44.2246740Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:25:44.2247416Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:25:44.2247928Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-12-04T09:25:44.2248265Z REENABLED_ISSUES= 2025-12-04T09:25:44.2248477Z DOCS= 2025-12-04T09:25:44.2248653Z SHLVL=1 2025-12-04T09:25:44.2248844Z MAX_JOBS=14 2025-12-04T09:25:44.2249060Z GITHUB_ACTOR_ID=475357 2025-12-04T09:25:44.2249381Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:25:44.2249746Z GITHUB_REF_NAME=main 2025-12-04T09:25:44.2250108Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:25:44.2250511Z GITHUB_JOB=test 2025-12-04T09:25:44.2250727Z NO_TEST_TIMEOUT=False 2025-12-04T09:25:44.2250963Z TD_DISTRIBUTED=False 2025-12-04T09:25:44.2251221Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:25:44.2251508Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:25:44.2251763Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:25:44.2252023Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:25:44.2252742Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:25:44.2253488Z GITHUB_BASE_REF= 2025-12-04T09:25:44.2253707Z INSTALLED_ACL= 2025-12-04T09:25:44.2254089Z ARTIFACTS_FILE_SUFFIX=test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167 2025-12-04T09:25:44.2254536Z CI=true 2025-12-04T09:25:44.2254748Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:25:44.2255049Z RUST_LOG=sccache::server=error 2025-12-04T09:25:44.2255312Z JOB_ID=57118183167 2025-12-04T09:25:44.2255530Z GITHUB_HEAD_REF= 2025-12-04T09:25:44.2255749Z GITHUB_ACTION_REF= 2025-12-04T09:25:44.2256024Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:25:44.2256361Z TEST_SHOWLOCALS=False 2025-12-04T09:25:44.2256604Z GITHUB_WORKFLOW=periodic 2025-12-04T09:25:44.2256858Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:25:44.2257444Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_ec04e73f-652e-4a69-b022-f618d5ca47fa 2025-12-04T09:25:44.2258034Z NO_TD=False 2025-12-04T09:25:44.2258254Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:25:44.2258551Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:25:44.2258977Z OLDPWD=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:25:44.2259470Z _=/usr/bin/env 2025-12-04T09:25:44.2259689Z + echo 'Testing pytorch' 2025-12-04T09:25:44.2259932Z Testing pytorch 2025-12-04T09:25:44.2260155Z + export LANG=C.UTF-8 2025-12-04T09:25:44.2260391Z + LANG=C.UTF-8 2025-12-04T09:25:44.2260597Z + PR_NUMBER= 2025-12-04T09:25:44.2260815Z + [[ default == \d\e\f\a\u\l\t ]] 2025-12-04T09:25:44.2261094Z + export CUDA_VISIBLE_DEVICES=0 2025-12-04T09:25:44.2261361Z + CUDA_VISIBLE_DEVICES=0 2025-12-04T09:25:44.2261620Z + export HIP_VISIBLE_DEVICES=0 2025-12-04T09:25:44.2261884Z + HIP_VISIBLE_DEVICES=0 2025-12-04T09:25:44.2262138Z + [[ default == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-12-04T09:25:44.2262432Z + [[ default == \s\l\o\w ]] 2025-12-04T09:25:44.2262824Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *slow-gradcheck* ]] 2025-12-04T09:25:44.2263293Z + export PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 2025-12-04T09:25:44.2263619Z + PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 2025-12-04T09:25:44.2263928Z + export PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 2025-12-04T09:25:44.2264251Z + PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 2025-12-04T09:25:44.2264640Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *cuda* ]] 2025-12-04T09:25:44.2265059Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:25:44.2265386Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:25:44.2265684Z + [[ default == *crossref* ]] 2025-12-04T09:25:44.2266041Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *rocm* ]] 2025-12-04T09:25:44.2266509Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *xpu* ]] 2025-12-04T09:25:44.2266997Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *-bazel-* ]] 2025-12-04T09:25:44.2267395Z + pip_install ninja==1.10.2 2025-12-04T09:25:44.2267741Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-12-04T09:25:44.2268280Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-12-04T09:25:44.8414382Z Collecting ninja==1.10.2 2025-12-04T09:25:44.8660812Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-12-04T09:25:44.8990834Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-12-04T09:25:45.3028485Z Installing collected packages: ninja 2025-12-04T09:25:45.3028945Z Attempting uninstall: ninja 2025-12-04T09:25:45.3036039Z Found existing installation: ninja 1.11.1.4 2025-12-04T09:25:45.3061308Z Uninstalling ninja-1.11.1.4: 2025-12-04T09:25:45.3186401Z Successfully uninstalled ninja-1.11.1.4 2025-12-04T09:25:45.3848197Z Successfully installed ninja-1.10.2 2025-12-04T09:25:45.4440163Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:25:45.4442224Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:25:45.4443590Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *aarch64* ]] 2025-12-04T09:25:45.4444321Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *asan* ]] 2025-12-04T09:25:45.4444993Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *-debug* ]] 2025-12-04T09:25:45.4445682Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *-bazel-* ]] 2025-12-04T09:25:45.4446548Z + echo 'We are not in debug mode: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck. Expect the assertion to pass' 2025-12-04T09:25:45.4447429Z We are not in debug mode: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck. Expect the assertion to pass 2025-12-04T09:25:45.4447946Z + cd test 2025-12-04T09:25:45.4448282Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-12-04T09:25:47.1102189Z + [[ default == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-12-04T09:25:47.1102563Z + [[ default == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-12-04T09:25:47.1102912Z + [[ default == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-12-04T09:25:47.1108292Z + DYNAMO_BENCHMARK_FLAGS=() 2025-12-04T09:25:47.1108616Z + [[ default == *pr_time_benchmarks* ]] 2025-12-04T09:25:47.1108929Z + [[ default == *dynamo_eager* ]] 2025-12-04T09:25:47.1109220Z + [[ default == *aot_eager* ]] 2025-12-04T09:25:47.1109496Z + [[ default == *aot_inductor* ]] 2025-12-04T09:25:47.1109790Z + [[ default == *max_autotune_inductor* ]] 2025-12-04T09:25:47.1110097Z + [[ default == *inductor* ]] 2025-12-04T09:25:47.1110371Z + [[ default == *dynamic* ]] 2025-12-04T09:25:47.1110631Z + [[ default == *cpu* ]] 2025-12-04T09:25:47.1110871Z + [[ default == *xpu* ]] 2025-12-04T09:25:47.1111158Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-12-04T09:25:47.1139899Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *libtorch* ]] 2025-12-04T09:25:47.1140445Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *-bazel-* ]] 2025-12-04T09:25:47.1143022Z + cd test 2025-12-04T09:25:47.1143901Z + python -c 'import torch; print(torch.__config__.show())' 2025-12-04T09:25:48.7850634Z PyTorch built with: 2025-12-04T09:25:48.7851649Z - GCC 11.4 2025-12-04T09:25:48.7852022Z - C++ Version: 201703 2025-12-04T09:25:48.7852635Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:25:48.7853330Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:25:48.7853761Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:25:48.7854092Z - LAPACK is enabled (usually provided by MKL) 2025-12-04T09:25:48.7854520Z - NNPACK is enabled 2025-12-04T09:25:48.7854857Z - CPU capability usage: AVX2 2025-12-04T09:25:48.7855170Z - CUDA Runtime 12.8 2025-12-04T09:25:48.7855520Z - NVCC architecture flags: -gencode;arch=compute_86,code=sm_86 2025-12-04T09:25:48.7856476Z - CuDNN 91.0.2 (built against CUDA 12.9) 2025-12-04T09:25:48.7862013Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32, CUDA_VERSION=12.8, CUDNN_VERSION=9.10.2, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, FORCE_FALLBACK_CUDA_MPI=1, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=ON, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=ON, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-12-04T09:25:48.7866704Z 2025-12-04T09:25:49.1625711Z + cd test 2025-12-04T09:25:49.1626091Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-12-04T09:25:50.4904193Z ATen/Parallel: 2025-12-04T09:25:50.4904519Z at::get_num_threads() : 8 2025-12-04T09:25:50.4904839Z at::get_num_interop_threads() : 16 2025-12-04T09:25:50.4905140Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:25:50.4905432Z omp_get_max_threads() : 8 2025-12-04T09:25:50.4905982Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:25:50.4906537Z mkl_get_max_threads() : 8 2025-12-04T09:25:50.4906917Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:25:50.4907384Z std::thread::hardware_concurrency() : 16 2025-12-04T09:25:50.4907692Z Environment variables: 2025-12-04T09:25:50.4907953Z OMP_NUM_THREADS : [not set] 2025-12-04T09:25:50.4908232Z MKL_NUM_THREADS : [not set] 2025-12-04T09:25:50.4908859Z ATen parallel backend: OpenMP 2025-12-04T09:25:50.4909049Z 2025-12-04T09:25:50.8187048Z + [[ default == *numpy_2* ]] 2025-12-04T09:25:50.8187653Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *aarch64* ]] 2025-12-04T09:25:50.8188225Z + [[ default == *backward* ]] 2025-12-04T09:25:50.8188641Z + [[ default == *libtorch_agnostic_targetting* ]] 2025-12-04T09:25:50.8189087Z + [[ default == *xla* ]] 2025-12-04T09:25:50.8189418Z + [[ default == *vllm* ]] 2025-12-04T09:25:50.8189757Z + [[ default == *executorch* ]] 2025-12-04T09:25:50.8190118Z + [[ default == \j\i\t\_\l\e\g\a\c\y ]] 2025-12-04T09:25:50.8190496Z + [[ default == \q\u\a\n\t\i\z\a\t\i\o\n ]] 2025-12-04T09:25:50.8190915Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *libtorch* ]] 2025-12-04T09:25:50.8191358Z + [[ default == distributed ]] 2025-12-04T09:25:50.8191652Z + [[ default == *operator_benchmark* ]] 2025-12-04T09:25:50.8191989Z + [[ default == *operator_microbenchmark* ]] 2025-12-04T09:25:50.8192348Z + [[ default == *attention_microbenchmark* ]] 2025-12-04T09:25:50.8192695Z + [[ default == *inductor_distributed* ]] 2025-12-04T09:25:50.8193017Z + [[ default == *inductor-halide* ]] 2025-12-04T09:25:50.8193317Z + [[ default == *inductor-pallas* ]] 2025-12-04T09:25:50.8193632Z + [[ default == *inductor-triton-cpu* ]] 2025-12-04T09:25:50.8193972Z + [[ default == *inductor-micro-benchmark* ]] 2025-12-04T09:25:50.8194337Z + [[ default == *aoti_cross_compile_for_windows* ]] 2025-12-04T09:25:50.8194674Z + [[ default == *huggingface* ]] 2025-12-04T09:25:50.8194956Z + [[ default == *timm* ]] 2025-12-04T09:25:50.8195218Z + [[ default == cachebench ]] 2025-12-04T09:25:50.8195502Z + [[ default == verify_cachebench ]] 2025-12-04T09:25:50.8195904Z + [[ default == *torchbench* ]] 2025-12-04T09:25:50.8196692Z + [[ default == *inductor_cpp_wrapper* ]] 2025-12-04T09:25:50.8197012Z + [[ default == *inductor_core* ]] 2025-12-04T09:25:50.8197304Z + [[ default == *inductor* ]] 2025-12-04T09:25:50.8197575Z + [[ default == *einops* ]] 2025-12-04T09:25:50.8197842Z + [[ default == *dynamo_core* ]] 2025-12-04T09:25:50.8198130Z + [[ default == *dynamo_wrapped* ]] 2025-12-04T09:25:50.8198530Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *rocm* ]] 2025-12-04T09:25:50.8198909Z + [[ 2 == 1 ]] 2025-12-04T09:25:50.8199113Z + [[ 2 == 2 ]] 2025-12-04T09:25:50.8199326Z + [[ 8 -gt 1 ]] 2025-12-04T09:25:50.8199549Z + install_torchvision 2025-12-04T09:25:50.8199989Z + local orig_preload 2025-12-04T09:25:50.8200222Z + local commit 2025-12-04T09:25:50.8200452Z ++ get_pinned_commit vision 2025-12-04T09:25:50.8200733Z ++ cat .github/ci_commit_pins/vision.txt 2025-12-04T09:25:50.8214336Z + commit=617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:25:50.8214795Z + orig_preload= 2025-12-04T09:25:50.8215073Z + '[' -n '' ']' 2025-12-04T09:25:50.8215529Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *cuda* ]] 2025-12-04T09:25:50.8216036Z + export FORCE_CUDA=1 2025-12-04T09:25:50.8216354Z + FORCE_CUDA=1 2025-12-04T09:25:50.8216649Z + export WITH_CUDA=1 2025-12-04T09:25:50.8216956Z + WITH_CUDA=1 2025-12-04T09:25:50.8217657Z + pip_build_and_install git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e dist/vision 2025-12-04T09:25:50.8218501Z + local build_target=git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:25:50.8219035Z + local wheel_dir=dist/vision 2025-12-04T09:25:50.8219298Z + local found_whl=0 2025-12-04T09:25:50.8219540Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:25:50.8219831Z + [[ -f dist/vision/*.whl ]] 2025-12-04T09:25:50.8229418Z + '[' 0 == 0 ']' 2025-12-04T09:25:50.8230157Z + python3 -m pip wheel --no-build-isolation --no-deps -w dist/vision git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:25:51.1456900Z Collecting git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:25:51.1461723Z Cloning https://github.com/pytorch/vision.git (to revision 617079d944b0e72632311c30ae2bbdf1168b901e) to /tmp/pip-req-build-a9e2okv8 2025-12-04T09:25:51.1609308Z Running command git clone --filter=blob:none --quiet https://github.com/pytorch/vision.git /tmp/pip-req-build-a9e2okv8 2025-12-04T09:25:52.8591445Z Running command git rev-parse -q --verify 'sha^617079d944b0e72632311c30ae2bbdf1168b901e' 2025-12-04T09:25:52.8618249Z Running command git fetch -q https://github.com/pytorch/vision.git 617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:25:52.9850357Z Resolved https://github.com/pytorch/vision.git to commit 617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:25:55.1136404Z Preparing metadata (pyproject.toml) ... [?25l- \ | done 2025-12-04T09:25:55.1173737Z [?25hBuilding wheels for collected packages: torchvision 2025-12-04T09:27:13.8972197Z Building wheel for torchvision (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | done 2025-12-04T09:27:13.9002491Z [?25h Created wheel for torchvision: filename=torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl size=1786465 sha256=39dde24eef4494d0388c2002db5130466e6af081dd160f0dab64fb2533e368c6 2025-12-04T09:27:13.9005648Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/12/b2/29/1f82685c5b5173629e1f36a9b93989ce92ce563e5fb91d27ac 2025-12-04T09:27:13.9044158Z Successfully built torchvision 2025-12-04T09:27:14.0161895Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:27:14.0162985Z + pip_install_whl dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:27:14.0163820Z + args=('dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl') 2025-12-04T09:27:14.0164270Z + local args 2025-12-04T09:27:14.0165034Z + [[ dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl == *\ * ]] 2025-12-04T09:27:14.0165592Z + for path in "${args[@]}" 2025-12-04T09:27:14.0166065Z + echo 'Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl' 2025-12-04T09:27:14.0166828Z Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:27:14.0167755Z + python3 -mpip install --no-index --no-deps dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:27:14.3511724Z Processing ./dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:27:14.3608445Z Installing collected packages: torchvision 2025-12-04T09:27:14.8240265Z Successfully installed torchvision-0.25.0a0+617079d 2025-12-04T09:27:14.8628638Z + '[' -n '' ']' 2025-12-04T09:27:14.8628925Z + test_python_shard 2 2025-12-04T09:27:14.8629169Z + [[ -z 8 ]] 2025-12-04T09:27:14.8629893Z + python test/run_test.py --exclude-jit-executor --exclude-distributed-tests --exclude-quantization-tests --shard 2 8 --verbose --upload-artifacts-while-running 2025-12-04T09:27:17.9668727Z Excluding doctests Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9669728Z Excluding test_meta Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9670532Z Excluding test_hub Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9671197Z Excluding test_fx Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9671881Z Excluding test_decomp Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9672705Z Excluding test_cpp_extensions_jit Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9673533Z Excluding test_jit Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9674316Z Excluding test_matmul_cuda Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9675081Z Excluding test_ops Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9676181Z Excluding test_ops_jit Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9677043Z Excluding dynamo/test_recompile_ux Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9677990Z Excluding inductor/test_compiled_optimizers Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9678967Z Excluding inductor/test_cutlass_backend Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9679892Z Excluding inductor/test_max_autotune Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9680824Z Excluding inductor/test_select_algorithm Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:17.9681738Z Excluding inductor/test_smoke Running in slow gradcheck mode, skipping tests that don't use gradcheck. 2025-12-04T09:27:19.9434922Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to /var/lib/jenkins/workspace/test/.pytorch-disabled-tests.json 2025-12-04T09:27:19.9983474Z Ignoring disabled issues: [''] 2025-12-04T09:27:20.0085417Z Found test times from artifacts 2025-12-04T09:27:20.0487780Z Found test times from artifacts 2025-12-04T09:27:20.0500282Z Running all tests 2025-12-04T09:27:20.1130273Z Running parallel tests on 1 processes 2025-12-04T09:27:20.1135971Z Name: tests to run (est. time: 140.51min) 2025-12-04T09:27:20.1136542Z Serial tests (80): 2025-12-04T09:27:20.1136924Z inductor/test_aot_inductor 2/5 2025-12-04T09:27:20.1137399Z inductor/test_torchinductor_codegen_dynamic_shapes 1/4 2025-12-04T09:27:20.1137880Z inductor/test_torchinductor_opinfo 4/14 2025-12-04T09:27:20.1138235Z inductor/test_torchinductor_opinfo 12/14 2025-12-04T09:27:20.1139060Z inductor/test_flex_attention 6/6 2025-12-04T09:27:20.1139459Z inductor/test_fp8 1/1 2025-12-04T09:27:20.1139821Z dynamo/test_model_output 1/1 2025-12-04T09:27:20.1140163Z inductor/test_triton_kernels 1/1 2025-12-04T09:27:20.1140494Z inductor/test_loop_ordering 1/1 2025-12-04T09:27:20.1140792Z export/test_serdes 1/1 2025-12-04T09:27:20.1141079Z inductor/test_scatter_optimization 1/1 2025-12-04T09:27:20.1141401Z inductor/test_padding 1/1 2025-12-04T09:27:20.1141684Z dynamo/test_callback 1/1 2025-12-04T09:27:20.1141966Z inductor/test_custom_op_autotune 1/1 2025-12-04T09:27:20.1142303Z test_cuda 1/1 2025-12-04T09:27:20.1142613Z test_sparse 1/1 2025-12-04T09:27:20.1142855Z test_ci_sanity_check_fail 1/1 2025-12-04T09:27:20.1143230Z test_ops_fwd_gradients 6/12 2025-12-04T09:27:20.1143566Z test_ops_gradients 2/10 2025-12-04T09:27:20.1143972Z test_ops_gradients 10/10 2025-12-04T09:27:20.1144278Z functorch/test_ops 3/6 2025-12-04T09:27:20.1144590Z dynamo/test_after_aot 1/1 2025-12-04T09:27:20.1144964Z inductor/test_snode_runtime 1/1 2025-12-04T09:27:20.1145377Z inductor/test_compiled_autograd 1/2 2025-12-04T09:27:20.1145773Z test_testing 1/1 2025-12-04T09:27:20.1146056Z inductor/test_autoheuristic 1/1 2025-12-04T09:27:20.1146361Z inductor/test_cutedsl_template 1/1 2025-12-04T09:27:20.1146693Z inductor/test_benchmark_fusion 1/1 2025-12-04T09:27:20.1147111Z inductor/test_remote_cache 1/1 2025-12-04T09:27:20.1147462Z inductor/test_coordinate_descent_tuner 1/1 2025-12-04T09:27:20.1147902Z inductor/test_inplace_padding 1/1 2025-12-04T09:27:20.1148329Z inductor/test_cudacodecache 1/1 2025-12-04T09:27:20.1148742Z inductor/test_minifier_utils 1/1 2025-12-04T09:27:20.1149148Z inductor/test_debug_trace 1/1 2025-12-04T09:27:20.1149542Z export/test_tree_utils 1/1 2025-12-04T09:27:20.1149831Z inductor/test_triton_wrapper 1/1 2025-12-04T09:27:20.1150158Z inductor/test_static_cuda_launcher 1/1 2025-12-04T09:27:20.1150504Z inductor/test_provenance_tracing 1/1 2025-12-04T09:27:20.1150832Z inductor/test_memory_planning 1/1 2025-12-04T09:27:20.1151136Z export/test_cpp_serdes 1/1 2025-12-04T09:27:20.1151427Z inductor/test_control_flow 2/4 2025-12-04T09:27:20.1151951Z test_sort_and_select 1/1 2025-12-04T09:27:20.1152233Z functorch/test_rearrange 1/1 2025-12-04T09:27:20.1152522Z test_package 1/1 2025-12-04T09:27:20.1152772Z test_mkl_verbose 1/1 2025-12-04T09:27:20.1153031Z test_utils_config_module 1/1 2025-12-04T09:27:20.1153312Z test_hop_infra 1/1 2025-12-04T09:27:20.1153582Z test_appending_byte_serializer 1/1 2025-12-04T09:27:20.1153881Z test_ao_sparsity 1/1 2025-12-04T09:27:20.1154145Z test_extension_utils 1/1 2025-12-04T09:27:20.1154426Z nn/attention/test_fa4 1/1 2025-12-04T09:27:20.1154711Z typing/test_python_operators 1/1 2025-12-04T09:27:20.1155019Z torch_np/test_dtype 1/1 2025-12-04T09:27:20.1155287Z test_file_check 1/1 2025-12-04T09:27:20.1155543Z profiler/test_kineto 1/1 2025-12-04T09:27:20.1155832Z functorch/test_ac_knapsack 1/1 2025-12-04T09:27:20.1156138Z torch_np/test_nep50_examples 1/1 2025-12-04T09:27:20.1156428Z test_torch 1/1 2025-12-04T09:27:20.1156660Z xpu/test_gemm 1/1 2025-12-04T09:27:20.1156919Z test_binary_ufuncs 1/1 2025-12-04T09:27:20.1157179Z test_modules 2/4 2025-12-04T09:27:20.1157450Z torch_np/numpy_tests/linalg/test_linalg 1/1 2025-12-04T09:27:20.1157809Z torch_np/numpy_tests/core/test_dtype 1/1 2025-12-04T09:27:20.1158127Z lazy/test_debug_util 1/1 2025-12-04T09:27:20.1158393Z nn/test_load_state_dict 1/1 2025-12-04T09:27:20.1158670Z test_shape_ops 1/1 2025-12-04T09:27:20.1158944Z profiler/test_memory_profiler 1/1 2025-12-04T09:27:20.1159236Z test_indexing 1/1 2025-12-04T09:27:20.1159480Z test_type_info 1/1 2025-12-04T09:27:20.1159750Z functorch/test_aotdispatch 1/1 2025-12-04T09:27:20.1160042Z test_scatter_gather_ops 1/1 2025-12-04T09:27:20.1160413Z test_cuda_multigpu 1/1 2025-12-04T09:27:20.1160725Z torch_np/numpy_tests/lib/test_index_tricks 1/1 2025-12-04T09:27:20.1161055Z test_jit_autocast 1/1 2025-12-04T09:27:20.1161326Z test_xnnpack_integration 1/1 2025-12-04T09:27:20.1161610Z nn/test_init 1/1 2025-12-04T09:27:20.1161859Z test_mobile_optimizer 1/1 2025-12-04T09:27:20.1162136Z test_type_promotion 1/1 2025-12-04T09:27:20.1162406Z test_reductions 1/1 2025-12-04T09:27:20.1162663Z test_autoload_disable 1/1 2025-12-04T09:27:20.1162941Z Parallel tests (0): 2025-12-04T09:27:20.1163194Z Name: excluded (est. time: 0.0min) 2025-12-04T09:27:20.1163479Z Serial tests (0): 2025-12-04T09:27:20.1163718Z Parallel tests (0): 2025-12-04T09:27:20.1164121Z Running inductor/test_aot_inductor 2/5 ... [2025-12-04 09:27:20.114173][942.5306447] 2025-12-04T09:27:20.1164569Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:27:20.1165737Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor.py', '--shard-id=2', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:27:20.114614] 2025-12-04T09:31:21.1097408Z 2025-12-04T09:31:21.1100882Z inductor/test_aot_inductor 2/5 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_2.5_0c57284626130d23_.log 2025-12-04T09:31:21.1123683Z Running 50 items in this shard: test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:31:21.1145240Z 2025-12-04T09:31:21.1145536Z Finished inductor/test_aot_inductor 2/5 ... [2025-12-04 09:31:21.108983][1183.525454214], took 4.02min 2025-12-04T09:31:21.1146566Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-7db6a6e521716bbc.xml 2025-12-04T09:31:21.5637575Z Uploading artifacts took 0.15 seconds 2025-12-04T09:31:21.5641866Z Running inductor/test_torchinductor_codegen_dynamic_shapes 1/4 ... [2025-12-04 09:31:21.563847][1183.980318707] 2025-12-04T09:31:21.5642437Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:21.5646577Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_codegen_dynamic_shapes.py', '--shard-id=1', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:21.564280] 2025-12-04T09:31:31.4478360Z 2025-12-04T09:31:31.4479962Z inductor/test_torchinductor_codegen_dynamic_shapes 1/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_codegen_dynamic_shapes_1.4_6c2fdc6084ab6b1c_.log 2025-12-04T09:31:31.4481249Z Running 0 items in this shard: 2025-12-04T09:31:31.4481440Z 2025-12-04T09:31:31.4481852Z Finished inductor/test_torchinductor_codegen_dynamic_shapes 1/4 ... [2025-12-04 09:31:31.447328][1193.863798621], took 0.16min 2025-12-04T09:31:31.4486941Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_codegen_dynamic_shapes/inductor.test_torchinductor_codegen_dynamic_shapes-667397d8b0e0d249.xml 2025-12-04T09:31:31.5190781Z Running inductor/test_torchinductor_opinfo 4/14 ... [2025-12-04 09:31:31.518605][1193.935076499] 2025-12-04T09:31:31.5192166Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:31.5195588Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '--shard-id=4', '--num-shards=14', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:31.518945] 2025-12-04T09:31:45.6094818Z 2025-12-04T09:31:45.6095932Z inductor/test_torchinductor_opinfo 4/14 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_4.14_349056dae69d6aaf_.log 2025-12-04T09:31:45.6119707Z Running 50 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float64 2025-12-04T09:31:45.6141818Z 2025-12-04T09:31:45.6142154Z Finished inductor/test_torchinductor_opinfo 4/14 ... [2025-12-04 09:31:45.609184][1208.025656989], took 0.23min 2025-12-04T09:31:45.6143310Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-274766fd55a1aebe.xml 2025-12-04T09:31:45.6915406Z Running inductor/test_torchinductor_opinfo 12/14 ... [2025-12-04 09:31:45.691086][1208.10755429] 2025-12-04T09:31:45.6916068Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:45.6920584Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '--shard-id=12', '--num-shards=14', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:45.691450] 2025-12-04T09:31:59.8803612Z 2025-12-04T09:31:59.8805701Z inductor/test_torchinductor_opinfo 12/14 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_12.14_d073bcb959e81001_.log 2025-12-04T09:31:59.8852965Z Running 100 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_float32 2025-12-04T09:31:59.8896130Z 2025-12-04T09:31:59.8896471Z Finished inductor/test_torchinductor_opinfo 12/14 ... [2025-12-04 09:31:59.880143][1222.296612603], took 0.24min 2025-12-04T09:31:59.8897615Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-deb335f8881dbf3d.xml 2025-12-04T09:31:59.9594853Z Running inductor/test_flex_attention 6/6 ... [2025-12-04 09:31:59.959143][1222.375615544] 2025-12-04T09:31:59.9595331Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:31:59.9598943Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_flex_attention.py', '--shard-id=6', '--num-shards=6', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:31:59.959471] 2025-12-04T09:33:45.0697130Z 2025-12-04T09:33:45.0698189Z inductor/test_flex_attention 6/6 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_flex_attention_6.6_b311f051d2033e5b_.log 2025-12-04T09:33:45.0752365Z Running 100 items in this shard: test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16 2025-12-04T09:33:45.0800662Z 2025-12-04T09:33:45.0800967Z Finished inductor/test_flex_attention 6/6 ... [2025-12-04 09:33:45.069648][1327.486120904], took 1.75min 2025-12-04T09:33:45.0802032Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_flex_attention/inductor.test_flex_attention-92d0889cf260b412.xml 2025-12-04T09:33:45.1514395Z Running inductor/test_fp8 1/1 ... [2025-12-04 09:33:45.151063][1327.567533933] 2025-12-04T09:33:45.1514841Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:45.1517678Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_fp8.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:45.151409] 2025-12-04T09:33:51.7775390Z 2025-12-04T09:33:51.7776294Z inductor/test_fp8 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_fp8_1.1_ce1db0883598be82_.log 2025-12-04T09:33:51.7776988Z Running 0 items in this shard: 2025-12-04T09:33:51.7777195Z 2025-12-04T09:33:51.7777446Z Finished inductor/test_fp8 1/1 ... [2025-12-04 09:33:51.777187][1334.193660698], took 0.11min 2025-12-04T09:33:51.7789805Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_fp8/inductor.test_fp8-944065b13a109cb5.xml 2025-12-04T09:33:51.8532901Z Running dynamo/test_model_output 1/1 ... [2025-12-04 09:33:51.852877][1334.269349757] 2025-12-04T09:33:51.8533552Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:51.8536238Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_model_output.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:51.853204] 2025-12-04T09:33:55.9250387Z 2025-12-04T09:33:55.9251692Z dynamo/test_model_output 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_model_output_1.1_bdadec5e5fc8d868_.log 2025-12-04T09:33:55.9267911Z Running 50 items in this shard: test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr, test/dynamo/test_model_output.py::TestHFPretrained::test_pretrained_non_const_attr 2025-12-04T09:33:55.9283579Z 2025-12-04T09:33:55.9283861Z Finished dynamo/test_model_output 1/1 ... [2025-12-04 09:33:55.924756][1338.341228839], took 0.07min 2025-12-04T09:33:55.9284858Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-55abd8e6b4532ae9.xml 2025-12-04T09:33:55.9553695Z Running inductor/test_triton_kernels 1/1 ... [2025-12-04 09:33:55.955043][1338.371516605] 2025-12-04T09:33:55.9554180Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:33:55.9557476Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_triton_kernels.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:55.955369] 2025-12-04T09:34:11.1483984Z 2025-12-04T09:34:11.1485058Z PRINTING LOG FILE of inductor/test_triton_kernels 1/1 (test/test-reports/inductor.test_triton_kernels_1.1_7dfa100f00a9f3af_.log) 2025-12-04T09:34:11.1486235Z Test results will be stored in test-reports/python-pytest/inductor.test_triton_kernels/inductor.test_triton_kernels-68c9c700698bff68.xml 2025-12-04T09:34:11.1486985Z ============================= test session starts ============================== 2025-12-04T09:34:11.1487600Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:34:11.1488403Z cachedir: .pytest_cache 2025-12-04T09:34:11.1488985Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:34:11.1489633Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:34:11.1489935Z configfile: pytest.ini 2025-12-04T09:34:11.1490534Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:34:11.1491170Z collecting ... collected 366 items 2025-12-04T09:34:11.1491524Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:34:11.1505450Z Running 50 items in this shard: test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass, test/inductor/test_triton_kernels.py::CustomOpTests::test_subclass 2025-12-04T09:34:11.1518375Z 2025-12-04T09:34:11.1518668Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [1.6237s] [ 2%] 2025-12-04T09:34:11.1519328Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass FAILED [0.2189s] [ 2%] 2025-12-04T09:34:11.1519967Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1150s] [ 2%] 2025-12-04T09:34:11.1520602Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1146s] [ 2%] 2025-12-04T09:34:11.1521235Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1128s] [ 2%] 2025-12-04T09:34:11.1521862Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1128s] [ 2%] 2025-12-04T09:34:11.1522497Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1129s] [ 2%] 2025-12-04T09:34:11.1523132Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1130s] [ 2%] 2025-12-04T09:34:11.1523768Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1126s] [ 2%] 2025-12-04T09:34:11.1524409Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1126s] [ 2%] 2025-12-04T09:34:11.1525046Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1152s] [ 2%] 2025-12-04T09:34:11.1525681Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1145s] [ 2%] 2025-12-04T09:34:11.1526317Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1130s] [ 2%] 2025-12-04T09:34:11.1526954Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.2756s] [ 2%] 2025-12-04T09:34:11.1527664Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1135s] [ 2%] 2025-12-04T09:34:11.1528301Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1132s] [ 2%] 2025-12-04T09:34:11.1528932Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1132s] [ 2%] 2025-12-04T09:34:11.1529575Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1133s] [ 2%] 2025-12-04T09:34:11.1532506Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1137s] [ 2%] 2025-12-04T09:34:11.1533259Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1131s] [ 2%] 2025-12-04T09:34:11.1533901Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1130s] [ 2%] 2025-12-04T09:34:11.1534535Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1133s] [ 2%] 2025-12-04T09:34:11.1535170Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1133s] [ 2%] 2025-12-04T09:34:11.1536109Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1130s] [ 2%] 2025-12-04T09:34:11.1536755Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1131s] [ 2%] 2025-12-04T09:34:11.1537393Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1131s] [ 2%] 2025-12-04T09:34:11.1538039Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1130s] [ 2%] 2025-12-04T09:34:11.1538670Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1127s] [ 2%] 2025-12-04T09:34:11.1539315Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1129s] [ 2%] 2025-12-04T09:34:11.1539953Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1127s] [ 2%] 2025-12-04T09:34:11.1540583Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1124s] [ 2%] 2025-12-04T09:34:11.1541218Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1127s] [ 2%] 2025-12-04T09:34:11.1541848Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1127s] [ 2%] 2025-12-04T09:34:11.1542486Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1124s] [ 2%] 2025-12-04T09:34:11.1543115Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1120s] [ 2%] 2025-12-04T09:34:11.1543905Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1123s] [ 2%] 2025-12-04T09:34:11.1544542Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1123s] [ 2%] 2025-12-04T09:34:11.1545185Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1126s] [ 2%] 2025-12-04T09:34:11.1545820Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1122s] [ 2%] 2025-12-04T09:34:11.1546457Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1120s] [ 2%] 2025-12-04T09:34:11.1547093Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1123s] [ 2%] 2025-12-04T09:34:11.1547723Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1126s] [ 2%] 2025-12-04T09:34:11.1548361Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1124s] [ 2%] 2025-12-04T09:34:11.1548999Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1120s] [ 2%] 2025-12-04T09:34:11.1549646Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1123s] [ 2%] 2025-12-04T09:34:11.1550275Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1127s] [ 2%] 2025-12-04T09:34:11.1550967Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1125s] [ 2%] 2025-12-04T09:34:11.1551609Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1124s] [ 2%] 2025-12-04T09:34:11.1552248Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1124s] [ 2%] 2025-12-04T09:34:11.1552874Z inductor/test_triton_kernels.py::CustomOpTests::test_subclass PASSED [0.1123s] [ 2%] 2025-12-04T09:34:11.1553253Z 2025-12-04T09:34:11.1553386Z =================================== FAILURES =================================== 2025-12-04T09:34:11.1553834Z _________________________ CustomOpTests.test_subclass __________________________ 2025-12-04T09:34:11.1554245Z Traceback (most recent call last): 2025-12-04T09:34:11.1554804Z File "/var/lib/jenkins/workspace/test/inductor/test_triton_kernels.py", line 3723, in test_subclass 2025-12-04T09:34:11.1555382Z self.assertEqual(out.a, expected.a) 2025-12-04T09:34:11.1556097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:34:11.1556717Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:34:11.1557396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:34:11.1558195Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:34:11.1558608Z AssertionError: Tensor-likes are not close! 2025-12-04T09:34:11.1558848Z 2025-12-04T09:34:11.1558959Z Mismatched elements: 3 / 3 (100.0%) 2025-12-04T09:34:11.1559410Z Greatest absolute difference: 3.620396614074707 at index (0,) (up to 1e-05 allowed) 2025-12-04T09:34:11.1560036Z Greatest relative difference: 2.2396645545959473 at index (0,) (up to 1.3e-06 allowed) 2025-12-04T09:34:11.1560430Z 2025-12-04T09:34:11.1560647Z To execute this test, run the following from the base repo dir: 2025-12-04T09:34:11.1561420Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_triton_kernels.py CustomOpTests.test_subclass 2025-12-04T09:34:11.1562011Z 2025-12-04T09:34:11.1562244Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:34:11.1562769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:34:11.1563195Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-12-04T09:34:11.1563551Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:34:11.1564016Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:34:11.1564468Z graph_break [] 2025-12-04T09:34:11.1564791Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:34:11.1565309Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-12-04T09:34:11.1565803Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:34:11.1566275Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T09:34:11.1566571Z graph_break [] 2025-12-04T09:34:11.1567424Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_triton_kernels/inductor.test_triton_kernels-68c9c700698bff68.xml - 2025-12-04T09:34:11.1568283Z =========================== short test summary info ============================ 2025-12-04T09:34:11.1568960Z FAILED [0.2189s] inductor/test_triton_kernels.py::CustomOpTests::test_subclass - AssertionError: Tensor-likes are not close! 2025-12-04T09:34:11.1569479Z 2025-12-04T09:34:11.1569585Z Mismatched elements: 3 / 3 (100.0%) 2025-12-04T09:34:11.1570032Z Greatest absolute difference: 3.620396614074707 at index (0,) (up to 1e-05 allowed) 2025-12-04T09:34:11.1570637Z Greatest relative difference: 2.2396645545959473 at index (0,) (up to 1.3e-06 allowed) 2025-12-04T09:34:11.1571015Z 2025-12-04T09:34:11.1571206Z To execute this test, run the following from the base repo dir: 2025-12-04T09:34:11.1571976Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_triton_kernels.py CustomOpTests.test_subclass 2025-12-04T09:34:11.1572547Z 2025-12-04T09:34:11.1572783Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:34:11.1573257Z ========================= 1 failed, 49 passed in 8.07s ========================= 2025-12-04T09:34:11.1573522Z 2025-12-04T09:34:11.1573978Z FINISHED PRINTING LOG FILE of inductor/test_triton_kernels 1/1 (test/test-reports/inductor.test_triton_kernels_1.1_7dfa100f00a9f3af_.log) 2025-12-04T09:34:11.1574548Z 2025-12-04T09:34:11.1574835Z Finished inductor/test_triton_kernels 1/1 ... [2025-12-04 09:34:11.148099][1353.564571948], took 0.25min 2025-12-04T09:34:11.1575881Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_triton_kernels/inductor.test_triton_kernels-68c9c700698bff68.xml 2025-12-04T09:34:11.4220131Z Uploading logs for 57118183167 to S3 2025-12-04T09:34:11.4495495Z Uploading artifacts took 0.22 seconds 2025-12-04T09:34:11.4495836Z inductor/test_triton_kernels 1/1 failed! 2025-12-04T09:34:11.4499443Z Running inductor/test_loop_ordering 1/1 ... [2025-12-04 09:34:11.449651][1353.866123877] 2025-12-04T09:34:11.4499920Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:11.4503490Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_loop_ordering.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:11.449986] 2025-12-04T09:34:17.8260021Z 2025-12-04T09:34:17.8260837Z inductor/test_loop_ordering 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_loop_ordering_1.1_2041ca8ae7154f7f_.log 2025-12-04T09:34:17.8279607Z Running 50 items in this shard: test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template, test/inductor/test_loop_ordering.py::LoopOrderingTest::test_interaction_with_triton_template 2025-12-04T09:34:17.8297612Z 2025-12-04T09:34:17.8297906Z Finished inductor/test_loop_ordering 1/1 ... [2025-12-04 09:34:17.825646][1360.242118032], took 0.11min 2025-12-04T09:34:17.8298951Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_loop_ordering/inductor.test_loop_ordering-5ee2a3a0a4d76c5b.xml 2025-12-04T09:34:17.9032356Z Running export/test_serdes 1/1 ... [2025-12-04 09:34:17.902881][1360.31935411] 2025-12-04T09:34:17.9032802Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:17.9036093Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_serdes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:17.903205] 2025-12-04T09:34:25.6313416Z 2025-12-04T09:34:25.6314362Z export/test_serdes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_serdes_1.1_6336b60dadd71d1f_.log 2025-12-04T09:34:25.6315060Z Running 0 items in this shard: 2025-12-04T09:34:25.6315262Z 2025-12-04T09:34:25.6315520Z Finished export/test_serdes 1/1 ... [2025-12-04 09:34:25.630955][1368.047425432], took 0.13min 2025-12-04T09:34:25.6336686Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_serdes/export.test_serdes-683bede13c6be09d.xml 2025-12-04T09:34:25.7039345Z Running inductor/test_scatter_optimization 1/1 ... [2025-12-04 09:34:25.703580][1368.120052466] 2025-12-04T09:34:25.7039856Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:25.7042716Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_scatter_optimization.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:25.703891] 2025-12-04T09:34:31.9294351Z 2025-12-04T09:34:31.9295483Z inductor/test_scatter_optimization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_scatter_optimization_1.1_34296d74f22943a6_.log 2025-12-04T09:34:31.9296391Z Running 0 items in this shard: 2025-12-04T09:34:31.9296615Z 2025-12-04T09:34:31.9296946Z Finished inductor/test_scatter_optimization 1/1 ... [2025-12-04 09:34:31.929108][1374.34558097], took 0.10min 2025-12-04T09:34:31.9317771Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_scatter_optimization/inductor.test_scatter_optimization-f067953bd7ad79f7.xml 2025-12-04T09:34:32.0109169Z Running inductor/test_padding 1/1 ... [2025-12-04 09:34:32.010543][1374.427015476] 2025-12-04T09:34:32.0109760Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:32.0113747Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_padding.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:32.010855] 2025-12-04T09:34:38.2859921Z 2025-12-04T09:34:38.2861207Z inductor/test_padding 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_padding_1.1_6c7d5d1307b4cd62_.log 2025-12-04T09:34:38.2861979Z Running 0 items in this shard: 2025-12-04T09:34:38.2862167Z 2025-12-04T09:34:38.2862458Z Finished inductor/test_padding 1/1 ... [2025-12-04 09:34:38.285627][1380.702100464], took 0.10min 2025-12-04T09:34:38.2885513Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-044d18a3f3486007.xml 2025-12-04T09:34:38.3570184Z Running dynamo/test_callback 1/1 ... [2025-12-04 09:34:38.356678][1380.773150982] 2025-12-04T09:34:38.3570645Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:38.3573968Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_callback.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:38.357034] 2025-12-04T09:34:44.5823046Z 2025-12-04T09:34:44.5823859Z dynamo/test_callback 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_callback_1.1_1fd2e27dc827492e_.log 2025-12-04T09:34:44.5824612Z Running 0 items in this shard: 2025-12-04T09:34:44.5824811Z 2025-12-04T09:34:44.5825075Z Finished dynamo/test_callback 1/1 ... [2025-12-04 09:34:44.581924][1386.998397585], took 0.10min 2025-12-04T09:34:44.5850597Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-3a59a6b9d0aa5ff8.xml 2025-12-04T09:34:44.6530177Z Running inductor/test_custom_op_autotune 1/1 ... [2025-12-04 09:34:44.652650][1387.069122432] 2025-12-04T09:34:44.6530800Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:44.6534338Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_custom_op_autotune.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:44.653015] 2025-12-04T09:34:50.8289302Z 2025-12-04T09:34:50.8290854Z inductor/test_custom_op_autotune 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_custom_op_autotune_1.1_d3a7b75a87bd527a_.log 2025-12-04T09:34:50.8291683Z Running 0 items in this shard: 2025-12-04T09:34:50.8291870Z 2025-12-04T09:34:50.8292186Z Finished inductor/test_custom_op_autotune 1/1 ... [2025-12-04 09:34:50.828524][1393.244997692], took 0.10min 2025-12-04T09:34:50.8317801Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_custom_op_autotune/inductor.test_custom_op_autotune-8053d2589810df17.xml 2025-12-04T09:34:50.9094619Z Running test_cuda 1/1 ... [2025-12-04 09:34:50.909068][1393.325539977] 2025-12-04T09:34:50.9095047Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:34:50.9098172Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_cuda.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:34:50.909413] 2025-12-04T09:36:41.1691643Z 2025-12-04T09:36:41.1692509Z PRINTING LOG FILE of test_cuda 1/1 (test/test-reports/test_cuda_1.1_2d01c4f5126b94de_.log) 2025-12-04T09:36:41.1693297Z Test results will be stored in test-reports/python-pytest/test_cuda/test_cuda-be50f111eb28e9fa.xml 2025-12-04T09:36:41.1694043Z ============================= test session starts ============================== 2025-12-04T09:36:41.1694815Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:36:41.1695438Z cachedir: .pytest_cache 2025-12-04T09:36:41.1696250Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:36:41.1697159Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:36:41.1697546Z configfile: pytest.ini 2025-12-04T09:36:41.1698183Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:36:41.1698873Z collecting ... collected 252 items 2025-12-04T09:36:41.1699319Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:36:41.1844266Z Running 550 items in this shard: test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_graph_concurrent_replay, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_host_memory_stats, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_invalid_status_for_legacy_api, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_record_stream_on_shifted_view, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_ctx_multithread, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator, test/test_cuda.py::TestMemPool::test_mempool_with_allocator 2025-12-04T09:36:41.1968181Z 2025-12-04T09:36:41.1968805Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0008s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 0%] 2025-12-04T09:36:41.1969878Z test_cuda.py::TestCuda::test_host_memory_stats PASSED [0.1467s] [ 0%] 2025-12-04T09:36:41.1970457Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0020s] [ 0%] 2025-12-04T09:36:41.1971408Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 0%] 2025-12-04T09:36:41.1972769Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 0%] 2025-12-04T09:36:41.1973773Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot FAILED [0.0601s] [ 1%] 2025-12-04T09:36:41.1974743Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp [W1204 09:34:55.063090515 unwind.cpp:219] Warning: Unsupported unwinding pattern: Address not in range (function unwinderFor) 2025-12-04T09:36:41.1975543Z FAILED [9.4466s] [ 1%] 2025-12-04T09:36:41.1975938Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread FAILED [0.0024s] [ 1%] 2025-12-04T09:36:41.1978386Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:36:41.1981153Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator.so 2025-12-04T09:36:41.1982043Z PASSED [16.9100s] [ 1%] 2025-12-04T09:36:41.1984154Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator_v1 -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:36:41.1986934Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator_v1.so 2025-12-04T09:36:41.1987824Z PASSED [16.6965s] [ 1%] 2025-12-04T09:36:41.1989995Z test_cuda.py::TestMemPool::test_mempool_with_allocator [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator_v2 -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:36:41.1992740Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator_v2.so 2025-12-04T09:36:41.1993633Z PASSED [16.4460s] [ 2%] 2025-12-04T09:36:41.1994390Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.1995648Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.1996906Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.1998164Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.1999418Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0008s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2000668Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2001908Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2003240Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2004504Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2005759Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2007004Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2008302Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0008s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2009554Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2010815Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2012068Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2014108Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2015360Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2016614Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2017952Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2019207Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2020463Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2021717Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2022968Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2024229Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2025483Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2044674Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2045940Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2047425Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2048670Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2049907Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2051148Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2052388Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2053642Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2054897Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2056154Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2057404Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2058659Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2060030Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2061294Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2062548Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2063798Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2065046Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2066298Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2067547Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2068797Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2070055Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2071351Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2072697Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2073958Z test_cuda.py::TestCuda::test_graph_concurrent_replay SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2074860Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1553s] [ 2%] 2025-12-04T09:36:41.2075408Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1540s] [ 2%] 2025-12-04T09:36:41.2075939Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1538s] [ 2%] 2025-12-04T09:36:41.2076480Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1538s] [ 2%] 2025-12-04T09:36:41.2077029Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1538s] [ 2%] 2025-12-04T09:36:41.2077560Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1541s] [ 2%] 2025-12-04T09:36:41.2078109Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1540s] [ 2%] 2025-12-04T09:36:41.2078645Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1550s] [ 2%] 2025-12-04T09:36:41.2079191Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1549s] [ 2%] 2025-12-04T09:36:41.2079719Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1547s] [ 2%] 2025-12-04T09:36:41.2080252Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1552s] [ 2%] 2025-12-04T09:36:41.2080799Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1553s] [ 2%] 2025-12-04T09:36:41.2081385Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1545s] [ 2%] 2025-12-04T09:36:41.2081913Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1550s] [ 2%] 2025-12-04T09:36:41.2082459Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1563s] [ 2%] 2025-12-04T09:36:41.2083088Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1548s] [ 2%] 2025-12-04T09:36:41.2083617Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1548s] [ 2%] 2025-12-04T09:36:41.2084150Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1543s] [ 2%] 2025-12-04T09:36:41.2084686Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1546s] [ 2%] 2025-12-04T09:36:41.2085226Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1554s] [ 2%] 2025-12-04T09:36:41.2085754Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1550s] [ 2%] 2025-12-04T09:36:41.2086290Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1555s] [ 2%] 2025-12-04T09:36:41.2086825Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1546s] [ 2%] 2025-12-04T09:36:41.2087407Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1543s] [ 2%] 2025-12-04T09:36:41.2087946Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1546s] [ 2%] 2025-12-04T09:36:41.2088488Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1546s] [ 2%] 2025-12-04T09:36:41.2089027Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1549s] [ 2%] 2025-12-04T09:36:41.2089560Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1547s] [ 2%] 2025-12-04T09:36:41.2090102Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1554s] [ 2%] 2025-12-04T09:36:41.2090643Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1547s] [ 2%] 2025-12-04T09:36:41.2091220Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1543s] [ 2%] 2025-12-04T09:36:41.2091764Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1547s] [ 2%] 2025-12-04T09:36:41.2092393Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1547s] [ 2%] 2025-12-04T09:36:41.2092931Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1549s] [ 2%] 2025-12-04T09:36:41.2093471Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1545s] [ 2%] 2025-12-04T09:36:41.2094017Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1554s] [ 2%] 2025-12-04T09:36:41.2094558Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1544s] [ 2%] 2025-12-04T09:36:41.2095102Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1546s] [ 2%] 2025-12-04T09:36:41.2095634Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1554s] [ 2%] 2025-12-04T09:36:41.2096175Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1547s] [ 2%] 2025-12-04T09:36:41.2096713Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1546s] [ 2%] 2025-12-04T09:36:41.2097248Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1549s] [ 2%] 2025-12-04T09:36:41.2097794Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1562s] [ 2%] 2025-12-04T09:36:41.2098336Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1554s] [ 2%] 2025-12-04T09:36:41.2098878Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1555s] [ 2%] 2025-12-04T09:36:41.2099415Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1549s] [ 2%] 2025-12-04T09:36:41.2099959Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1549s] [ 2%] 2025-12-04T09:36:41.2100496Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1548s] [ 2%] 2025-12-04T09:36:41.2101031Z test_cuda.py::TestCuda::test_host_memory_stats FAILED [0.1551s] [ 2%] 2025-12-04T09:36:41.2101596Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0018s] [ 2%] 2025-12-04T09:36:41.2102181Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2102767Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2103344Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2104004Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2104589Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2105164Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0015s] [ 2%] 2025-12-04T09:36:41.2105739Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2106325Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2106895Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2107462Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0015s] [ 2%] 2025-12-04T09:36:41.2108039Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2108614Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2109196Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2109765Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2110336Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0015s] [ 2%] 2025-12-04T09:36:41.2110944Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2111538Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2112118Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2112697Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2113360Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2113931Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0015s] [ 2%] 2025-12-04T09:36:41.2114522Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2115103Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2115679Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0024s] [ 2%] 2025-12-04T09:36:41.2116254Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2116835Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2117417Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0016s] [ 2%] 2025-12-04T09:36:41.2117988Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2118568Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2119168Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2119757Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2120347Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2120986Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0016s] [ 2%] 2025-12-04T09:36:41.2121565Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2122142Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2122713Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2123300Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2123879Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2124466Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2125248Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0016s] [ 2%] 2025-12-04T09:36:41.2125827Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0016s] [ 2%] 2025-12-04T09:36:41.2126405Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2126976Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2127599Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2128177Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2128747Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0016s] [ 2%] 2025-12-04T09:36:41.2129335Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0013s] [ 2%] 2025-12-04T09:36:41.2129922Z test_cuda.py::TestCuda::test_invalid_status_for_legacy_api PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2130863Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2132190Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2133463Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2134744Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2136485Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2137780Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2139051Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2140332Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2141622Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2142909Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2144197Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2145468Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2146746Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2148032Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2149479Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2150782Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2152112Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2153396Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2154680Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2155978Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2157277Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2158563Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2159851Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2161250Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2162601Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2163891Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2165160Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2166457Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2167818Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2169118Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2170415Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2171692Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2172978Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2174364Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2175651Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2176944Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2178223Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2179515Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2180807Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2182104Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2183391Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2184665Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2185944Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2187309Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2188603Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2189883Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2191171Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2192454Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2193742Z test_cuda.py::TestCuda::test_record_stream_on_shifted_view SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2195097Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2196510Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2197906Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2199404Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2200807Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2202271Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2203670Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2205076Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2206488Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2207942Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2209356Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2210755Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2212311Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2213716Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2215130Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2216539Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2217950Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2219368Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2220761Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2222209Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2223614Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2225106Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2226510Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2227910Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2229316Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2230754Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2232161Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2233570Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2234980Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2236694Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2238241Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2239641Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2241047Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2242493Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2243901Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2245303Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2246711Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2248182Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2249601Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2251125Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2252548Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2253962Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2255372Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2256801Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2258220Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2259639Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2261104Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2262633Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2264049Z test_cuda.py::TestCuda::test_repeat_graph_capture_cublas_workspace_memory SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:36:41.2265041Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0388s] [ 2%] 2025-12-04T09:36:41.2265643Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0658s] [ 2%] 2025-12-04T09:36:41.2266237Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0375s] [ 2%] 2025-12-04T09:36:41.2266842Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0370s] [ 2%] 2025-12-04T09:36:41.2267424Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0380s] [ 2%] 2025-12-04T09:36:41.2268019Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0370s] [ 2%] 2025-12-04T09:36:41.2268621Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0375s] [ 2%] 2025-12-04T09:36:41.2269211Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0373s] [ 2%] 2025-12-04T09:36:41.2269788Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0374s] [ 2%] 2025-12-04T09:36:41.2270379Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0374s] [ 2%] 2025-12-04T09:36:41.2270980Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0373s] [ 2%] 2025-12-04T09:36:41.2271569Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0373s] [ 2%] 2025-12-04T09:36:41.2272214Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0371s] [ 2%] 2025-12-04T09:36:41.2272809Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2273395Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0375s] [ 2%] 2025-12-04T09:36:41.2273992Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2274574Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0375s] [ 2%] 2025-12-04T09:36:41.2275248Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2275827Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0371s] [ 2%] 2025-12-04T09:36:41.2276408Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2276988Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2277575Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0369s] [ 2%] 2025-12-04T09:36:41.2278155Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0377s] [ 2%] 2025-12-04T09:36:41.2278740Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0374s] [ 2%] 2025-12-04T09:36:41.2279329Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0373s] [ 2%] 2025-12-04T09:36:41.2279909Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0368s] [ 2%] 2025-12-04T09:36:41.2280503Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2281085Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0371s] [ 2%] 2025-12-04T09:36:41.2281662Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0370s] [ 2%] 2025-12-04T09:36:41.2282236Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0373s] [ 2%] 2025-12-04T09:36:41.2282826Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0369s] [ 2%] 2025-12-04T09:36:41.2283408Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0368s] [ 2%] 2025-12-04T09:36:41.2283991Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0373s] [ 2%] 2025-12-04T09:36:41.2284572Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2285249Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2285837Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0370s] [ 2%] 2025-12-04T09:36:41.2286421Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0372s] [ 2%] 2025-12-04T09:36:41.2287003Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0375s] [ 2%] 2025-12-04T09:36:41.2287620Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0375s] [ 2%] 2025-12-04T09:36:41.2288202Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0373s] [ 2%] 2025-12-04T09:36:41.2288767Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0367s] [ 2%] 2025-12-04T09:36:41.2289345Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0370s] [ 2%] 2025-12-04T09:36:41.2289924Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0371s] [ 2%] 2025-12-04T09:36:41.2290511Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0371s] [ 2%] 2025-12-04T09:36:41.2291138Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0373s] [ 2%] 2025-12-04T09:36:41.2291725Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0375s] [ 2%] 2025-12-04T09:36:41.2292299Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0381s] [ 2%] 2025-12-04T09:36:41.2292867Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0374s] [ 2%] 2025-12-04T09:36:41.2293441Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot PASSED [0.0379s] [ 2%] 2025-12-04T09:36:41.2294050Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0016s] [ 2%] 2025-12-04T09:36:41.2294691Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2295331Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2295975Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2296609Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2297323Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2297961Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0015s] [ 2%] 2025-12-04T09:36:41.2298599Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2299235Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2299863Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2300495Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0015s] [ 2%] 2025-12-04T09:36:41.2301125Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2301764Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2302402Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2303037Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2303668Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2304292Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2304925Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2305560Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2306200Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0015s] [ 2%] 2025-12-04T09:36:41.2306827Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2307566Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2308209Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2308838Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2309464Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2310091Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2310721Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2311397Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2312026Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2312663Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2313291Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2313927Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2314558Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2315188Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2315816Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2316444Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2317075Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2317708Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2318342Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2318973Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2319685Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2320312Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2320937Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2321616Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2322243Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2322867Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2323496Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2324134Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2324769Z test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp PASSED [0.0014s] [ 2%] 2025-12-04T09:36:41.2325361Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0025s] [ 2%] 2025-12-04T09:36:41.2325923Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2326479Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0021s] [ 2%] 2025-12-04T09:36:41.2327034Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2327668Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2328225Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2328793Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2329451Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2330025Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0021s] [ 2%] 2025-12-04T09:36:41.2330605Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2331181Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0022s] [ 2%] 2025-12-04T09:36:41.2331747Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2332324Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2332896Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2333458Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2334026Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0022s] [ 2%] 2025-12-04T09:36:41.2334603Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2335169Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2336039Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2336607Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0021s] [ 2%] 2025-12-04T09:36:41.2337175Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2337748Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2338307Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2338885Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2339462Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2340022Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2340616Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2341351Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2341917Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2342478Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0022s] [ 2%] 2025-12-04T09:36:41.2343046Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2343616Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2344183Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2344760Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2345341Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2345930Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2346496Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2347080Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2347651Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2348222Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2348792Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2349370Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0021s] [ 2%] 2025-12-04T09:36:41.2349949Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2350504Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2351176Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2351736Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2352300Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2352853Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2353421Z test_cuda.py::TestMemPool::test_mempool_ctx_multithread PASSED [0.0020s] [ 2%] 2025-12-04T09:36:41.2354003Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0017s] [ 2%] 2025-12-04T09:36:41.2354595Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2355180Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2362777Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2363384Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2363971Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2364556Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2365136Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2365707Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2366286Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2366866Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2367522Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2368104Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2368692Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2369281Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2369972Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2370552Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2371138Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2371773Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2372354Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2372946Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2373532Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2374112Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2374698Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2375286Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2375873Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2376465Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2377046Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2377631Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2378213Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2378786Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2379370Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2380038Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2380627Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0016s] [ 2%] 2025-12-04T09:36:41.2381210Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2381799Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2382386Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2382961Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2383547Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2384136Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2384721Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0019s] [ 2%] 2025-12-04T09:36:41.2385310Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2385905Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2386489Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2387082Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2387663Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2388252Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2388840Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0013s] [ 2%] 2025-12-04T09:36:41.2389419Z test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive FAILED [0.0014s] [ 2%] 2025-12-04T09:36:41.2391892Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator_v3 -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:36:41.2394666Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator_v3.so 2025-12-04T09:36:41.2395561Z PASSED [16.5445s] [ 2%] 2025-12-04T09:36:41.2396007Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0231s] [ 2%] 2025-12-04T09:36:41.2396662Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0227s] [ 2%] 2025-12-04T09:36:41.2397330Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0233s] [ 2%] 2025-12-04T09:36:41.2397994Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2398644Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0229s] [ 2%] 2025-12-04T09:36:41.2399289Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0232s] [ 2%] 2025-12-04T09:36:41.2399942Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2400597Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0229s] [ 2%] 2025-12-04T09:36:41.2401297Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0231s] [ 2%] 2025-12-04T09:36:41.2401943Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2402692Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2403349Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2404012Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0227s] [ 2%] 2025-12-04T09:36:41.2404664Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2405313Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0229s] [ 2%] 2025-12-04T09:36:41.2405969Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0229s] [ 2%] 2025-12-04T09:36:41.2406616Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0227s] [ 2%] 2025-12-04T09:36:41.2407354Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0236s] [ 2%] 2025-12-04T09:36:41.2408016Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0227s] [ 2%] 2025-12-04T09:36:41.2408668Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2409324Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0229s] [ 2%] 2025-12-04T09:36:41.2409977Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2410627Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0225s] [ 2%] 2025-12-04T09:36:41.2411281Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2411926Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0227s] [ 2%] 2025-12-04T09:36:41.2412574Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0227s] [ 2%] 2025-12-04T09:36:41.2413226Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0231s] [ 2%] 2025-12-04T09:36:41.2413879Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2414618Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0229s] [ 2%] 2025-12-04T09:36:41.2415273Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2415922Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0225s] [ 2%] 2025-12-04T09:36:41.2416570Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0227s] [ 2%] 2025-12-04T09:36:41.2417231Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2417886Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0224s] [ 2%] 2025-12-04T09:36:41.2418544Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2419196Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0226s] [ 2%] 2025-12-04T09:36:41.2419854Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2420513Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2421169Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0224s] [ 2%] 2025-12-04T09:36:41.2421824Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0228s] [ 2%] 2025-12-04T09:36:41.2422481Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0229s] [ 2%] 2025-12-04T09:36:41.2423138Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2423786Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2424534Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0227s] [ 2%] 2025-12-04T09:36:41.2425186Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2425840Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0230s] [ 2%] 2025-12-04T09:36:41.2426493Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0225s] [ 2%] 2025-12-04T09:36:41.2427154Z test_cuda.py::TestMemPool::test_mempool_limited_memory_with_allocator PASSED [0.0229s] [ 2%] 2025-12-04T09:36:41.2429435Z test_cuda.py::TestMemPool::test_mempool_with_allocator [1/2] c++ -MMD -MF main.o.d -DTORCH_EXTENSION_NAME=dummy_allocator_v4 -DTORCH_API_INCLUDE_EXTENSION_H -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -isystem /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -isystem /usr/local/cuda/include -isystem /opt/conda/envs/py_3.10/include/python3.10 -fPIC -std=c++17 -c /var/lib/jenkins/.cache/torch_extensions/py310_cu128/dummy_allocator/main.cpp -o main.o 2025-12-04T09:36:41.2432164Z [2/2] c++ main.o -shared -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -lc10 -lc10_cuda -ltorch_cpu -ltorch_cuda -ltorch -ltorch_python -L/usr/local/cuda/lib64 -lcudart -o dummy_allocator_v4.so 2025-12-04T09:36:41.2433057Z PASSED [16.5168s] [ 2%] 2025-12-04T09:36:41.2433441Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0019s] [ 2%] 2025-12-04T09:36:41.2434011Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0016s] [ 2%] 2025-12-04T09:36:41.2434561Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0019s] [ 2%] 2025-12-04T09:36:41.2435113Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2436019Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0016s] [ 2%] 2025-12-04T09:36:41.2436572Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2437135Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2437819Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0018s] [ 2%] 2025-12-04T09:36:41.2438378Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0016s] [ 2%] 2025-12-04T09:36:41.2438931Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2439492Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2440040Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2440591Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2441147Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2441743Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2442300Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2442853Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2443415Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0016s] [ 2%] 2025-12-04T09:36:41.2443967Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0017s] [ 2%] 2025-12-04T09:36:41.2444527Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2445082Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0016s] [ 2%] 2025-12-04T09:36:41.2445639Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2446187Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2446737Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0017s] [ 2%] 2025-12-04T09:36:41.2447337Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2448000Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2448561Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2449111Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2449672Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0017s] [ 2%] 2025-12-04T09:36:41.2450223Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2450777Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2451337Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2451888Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0018s] [ 2%] 2025-12-04T09:36:41.2452445Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2453003Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2453560Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2454115Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2454677Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2455234Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0016s] [ 2%] 2025-12-04T09:36:41.2455791Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0018s] [ 2%] 2025-12-04T09:36:41.2456341Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2456893Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0016s] [ 2%] 2025-12-04T09:36:41.2457444Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2457997Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2458559Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0017s] [ 2%] 2025-12-04T09:36:41.2459210Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2459770Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2460323Z test_cuda.py::TestMemPool::test_mempool_with_allocator FAILED [0.0015s] [ 2%] 2025-12-04T09:36:41.2460656Z 2025-12-04T09:36:41.2460777Z =================================== FAILURES =================================== 2025-12-04T09:36:41.2461227Z ___________________ TestCudaMallocAsync.test_memory_snapshot ___________________ 2025-12-04T09:36:41.2461645Z Traceback (most recent call last): 2025-12-04T09:36:41.2462125Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 4091, in test_memory_snapshot 2025-12-04T09:36:41.2462647Z torch.cuda.memory._save_segment_usage(f.name) 2025-12-04T09:36:41.2463274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/memory.py", line 1288, in _save_segment_usage 2025-12-04T09:36:41.2463857Z f.write(_segments(snapshot)) 2025-12-04T09:36:41.2464401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/_memory_viz.py", line 158, in segments 2025-12-04T09:36:41.2464959Z return format_flamegraph(f.getvalue()) 2025-12-04T09:36:41.2465541Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/cuda/_memory_viz.py", line 100, in format_flamegraph 2025-12-04T09:36:41.2466225Z with tempfile.NamedTemporaryFile(mode="wb", suffix=".pl") as f: 2025-12-04T09:36:41.2466758Z File "/opt/conda/envs/py_3.10/lib/python3.10/tempfile.py", line 518, in __exit__ 2025-12-04T09:36:41.2467178Z self.close() 2025-12-04T09:36:41.2467525Z File "/opt/conda/envs/py_3.10/lib/python3.10/tempfile.py", line 525, in close 2025-12-04T09:36:41.2467945Z self._closer.close() 2025-12-04T09:36:41.2468406Z File "/opt/conda/envs/py_3.10/lib/python3.10/tempfile.py", line 462, in close 2025-12-04T09:36:41.2468805Z unlink(self.name) 2025-12-04T09:36:41.2469188Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpkd1nxy_c.pl' 2025-12-04T09:36:41.2469713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2470206Z Downloading flamegraph.pl to: /var/lib/jenkins/.cache//flamegraph.pl 2025-12-04T09:36:41.2470739Z ______________ TestCudaMallocAsync.test_memory_snapshot_with_cpp _______________ 2025-12-04T09:36:41.2471204Z Traceback (most recent call last): 2025-12-04T09:36:41.2471727Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 4131, in test_memory_snapshot_with_cpp 2025-12-04T09:36:41.2472270Z self.assertTrue("::rand" in str(b["frames"])) 2025-12-04T09:36:41.2472763Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 687, in assertTrue 2025-12-04T09:36:41.2473248Z raise self.failureException(msg) 2025-12-04T09:36:41.2473567Z AssertionError: False is not true 2025-12-04T09:36:41.2473756Z 2025-12-04T09:36:41.2473936Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2474691Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCudaMallocAsync.test_memory_snapshot_with_cpp 2025-12-04T09:36:41.2475262Z 2025-12-04T09:36:41.2475482Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2476015Z ___________________ TestMemPool.test_mempool_ctx_multithread ___________________ 2025-12-04T09:36:41.2476425Z Traceback (most recent call last): 2025-12-04T09:36:41.2476915Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5905, in test_mempool_ctx_multithread 2025-12-04T09:36:41.2477538Z self.assertEqual(len(segments), 0, "Expected empty pool in the beginning") 2025-12-04T09:36:41.2478294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2478997Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2479391Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2479601Z 2025-12-04T09:36:41.2479804Z Expected 0 but got 1. 2025-12-04T09:36:41.2480060Z Absolute difference: 1 2025-12-04T09:36:41.2480305Z Relative difference: inf 2025-12-04T09:36:41.2480573Z Expected empty pool in the beginning 2025-12-04T09:36:41.2480771Z 2025-12-04T09:36:41.2480980Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2481720Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_ctx_multithread 2025-12-04T09:36:41.2482256Z 2025-12-04T09:36:41.2482481Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2482993Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2483391Z Traceback (most recent call last): 2025-12-04T09:36:41.2483872Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2484342Z check_stats(expected) 2025-12-04T09:36:41.2484750Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2485191Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2485801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2486496Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2486883Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2487090Z 2025-12-04T09:36:41.2487186Z Expected 196608 but got 0. 2025-12-04T09:36:41.2487491Z Absolute difference: 196608 2025-12-04T09:36:41.2487751Z Relative difference: 1.0 2025-12-04T09:36:41.2487914Z 2025-12-04T09:36:41.2488095Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2488955Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2489460Z 2025-12-04T09:36:41.2489684Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2490182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2490612Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2491041Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2491433Z Traceback (most recent call last): 2025-12-04T09:36:41.2491908Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2492382Z check_stats(expected) 2025-12-04T09:36:41.2492774Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2493218Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2493835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2494528Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2494919Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2495127Z 2025-12-04T09:36:41.2495224Z Expected 196608 but got 0. 2025-12-04T09:36:41.2495481Z Absolute difference: 196608 2025-12-04T09:36:41.2495739Z Relative difference: 1.0 2025-12-04T09:36:41.2495906Z 2025-12-04T09:36:41.2496084Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2496765Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2497268Z 2025-12-04T09:36:41.2497490Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2497983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2498428Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2498850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2499354Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2499785Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2500173Z Traceback (most recent call last): 2025-12-04T09:36:41.2500640Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2501166Z check_stats(expected) 2025-12-04T09:36:41.2501567Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2502012Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2502616Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2503310Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2503709Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2503914Z 2025-12-04T09:36:41.2504007Z Expected 196608 but got 0. 2025-12-04T09:36:41.2504263Z Absolute difference: 196608 2025-12-04T09:36:41.2504536Z Relative difference: 1.0 2025-12-04T09:36:41.2504700Z 2025-12-04T09:36:41.2504883Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2505564Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2506070Z 2025-12-04T09:36:41.2506288Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2506786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2507219Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2507635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2508155Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2508567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2508990Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2509430Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2509818Z Traceback (most recent call last): 2025-12-04T09:36:41.2510289Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2510811Z check_stats(expected) 2025-12-04T09:36:41.2511206Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2511648Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2512250Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2512941Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2513331Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2513539Z 2025-12-04T09:36:41.2513633Z Expected 196608 but got 0. 2025-12-04T09:36:41.2513885Z Absolute difference: 196608 2025-12-04T09:36:41.2514155Z Relative difference: 1.0 2025-12-04T09:36:41.2514315Z 2025-12-04T09:36:41.2514495Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2515168Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2515669Z 2025-12-04T09:36:41.2515885Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2516377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2516807Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2517221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2517654Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2518069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2518492Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2519045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2519479Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2519905Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2520289Z Traceback (most recent call last): 2025-12-04T09:36:41.2520767Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2521240Z check_stats(expected) 2025-12-04T09:36:41.2521631Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2522069Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2522676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2523372Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2523757Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2523970Z 2025-12-04T09:36:41.2524060Z Expected 196608 but got 0. 2025-12-04T09:36:41.2524320Z Absolute difference: 196608 2025-12-04T09:36:41.2524576Z Relative difference: 1.0 2025-12-04T09:36:41.2524740Z 2025-12-04T09:36:41.2524921Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2525599Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2526100Z 2025-12-04T09:36:41.2526322Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2526813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2527241Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2527791Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2528214Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2528635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2529058Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2529470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2529890Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2530302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2530765Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2531194Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2531582Z Traceback (most recent call last): 2025-12-04T09:36:41.2532056Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2532532Z check_stats(expected) 2025-12-04T09:36:41.2532923Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2533367Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2533971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2534659Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2535046Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2535497Z 2025-12-04T09:36:41.2535594Z Expected 196608 but got 0. 2025-12-04T09:36:41.2535846Z Absolute difference: 196608 2025-12-04T09:36:41.2536103Z Relative difference: 1.0 2025-12-04T09:36:41.2536267Z 2025-12-04T09:36:41.2536446Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2537121Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2537632Z 2025-12-04T09:36:41.2537853Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2538479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2538909Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2539327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2539747Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2540162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2540586Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2541018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2541468Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2541883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2542308Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2542723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2543144Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2543571Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2543956Z Traceback (most recent call last): 2025-12-04T09:36:41.2544423Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2544901Z check_stats(expected) 2025-12-04T09:36:41.2545298Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2545739Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2546351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2547193Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2547576Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2547784Z 2025-12-04T09:36:41.2547885Z Expected 196608 but got 0. 2025-12-04T09:36:41.2548142Z Absolute difference: 196608 2025-12-04T09:36:41.2548398Z Relative difference: 1.0 2025-12-04T09:36:41.2548564Z 2025-12-04T09:36:41.2548743Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2549418Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2549924Z 2025-12-04T09:36:41.2550145Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2550644Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2551078Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2551507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2551933Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2552351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2552777Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2553200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2553624Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2554040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2554464Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2554887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2555308Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2555723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2556152Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2556574Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2557055Z Traceback (most recent call last): 2025-12-04T09:36:41.2557529Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2558001Z check_stats(expected) 2025-12-04T09:36:41.2558398Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2558845Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2559449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2560137Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2560518Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2560722Z 2025-12-04T09:36:41.2560836Z Expected 196608 but got 0. 2025-12-04T09:36:41.2561128Z Absolute difference: 196608 2025-12-04T09:36:41.2561390Z Relative difference: 1.0 2025-12-04T09:36:41.2561553Z 2025-12-04T09:36:41.2561730Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2562410Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2562912Z 2025-12-04T09:36:41.2563130Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2563626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2564052Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2564469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2564890Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2565303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2565821Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2566241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2584881Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2585327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2585773Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2586194Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2586621Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2587038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2587472Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2587886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2588314Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2588758Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2589149Z Traceback (most recent call last): 2025-12-04T09:36:41.2589621Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2590100Z check_stats(expected) 2025-12-04T09:36:41.2590510Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2590986Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2591604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2592298Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2592683Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2592888Z 2025-12-04T09:36:41.2592982Z Expected 196608 but got 0. 2025-12-04T09:36:41.2593244Z Absolute difference: 196608 2025-12-04T09:36:41.2593523Z Relative difference: 1.0 2025-12-04T09:36:41.2593685Z 2025-12-04T09:36:41.2593864Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2595191Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2595703Z 2025-12-04T09:36:41.2595928Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2596433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2596863Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2597287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2597715Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2598128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2598556Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2598980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2599413Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2599829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2600259Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2600722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2601140Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2601562Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2601988Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2602403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2602830Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2603361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2603790Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2604215Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2604615Z Traceback (most recent call last): 2025-12-04T09:36:41.2605092Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2605570Z check_stats(expected) 2025-12-04T09:36:41.2605961Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2606407Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2607020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2607768Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2608155Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2608382Z 2025-12-04T09:36:41.2608473Z Expected 196608 but got 0. 2025-12-04T09:36:41.2608735Z Absolute difference: 196608 2025-12-04T09:36:41.2608996Z Relative difference: 1.0 2025-12-04T09:36:41.2609165Z 2025-12-04T09:36:41.2609347Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2610034Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2610537Z 2025-12-04T09:36:41.2610757Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2611257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2611720Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2612171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2612595Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2613016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2613455Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2613954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2614389Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2614812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2615244Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2615659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2616090Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2616506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2616925Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2617344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2617780Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2618194Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2618621Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2619037Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2619463Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2619889Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2620280Z Traceback (most recent call last): 2025-12-04T09:36:41.2620750Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2621236Z check_stats(expected) 2025-12-04T09:36:41.2621639Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2622089Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2622778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2623466Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2623850Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2624052Z 2025-12-04T09:36:41.2624145Z Expected 196608 but got 0. 2025-12-04T09:36:41.2624404Z Absolute difference: 196608 2025-12-04T09:36:41.2624668Z Relative difference: 1.0 2025-12-04T09:36:41.2624832Z 2025-12-04T09:36:41.2625016Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2625690Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2626191Z 2025-12-04T09:36:41.2626413Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2626904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2627338Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2627759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2628187Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2628605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2629029Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2629439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2629882Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2630316Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2630777Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2631205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2631632Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2632065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2632492Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2632997Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2633431Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2633846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2634270Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2634693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2635123Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2635709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2636136Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2636575Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2636976Z Traceback (most recent call last): 2025-12-04T09:36:41.2637453Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2637931Z check_stats(expected) 2025-12-04T09:36:41.2638339Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2638781Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2639395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2640092Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2640479Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2640698Z 2025-12-04T09:36:41.2640799Z Expected 196608 but got 0. 2025-12-04T09:36:41.2641083Z Absolute difference: 196608 2025-12-04T09:36:41.2641354Z Relative difference: 1.0 2025-12-04T09:36:41.2641659Z 2025-12-04T09:36:41.2641840Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2642534Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2643046Z 2025-12-04T09:36:41.2643264Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2643767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2644199Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2644616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2645044Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2645461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2645887Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2646314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2646740Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2647164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2647639Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2648059Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2648482Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2648902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2649338Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2649762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2650182Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2650605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2651060Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2651508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2652060Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2652476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2652904Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2653317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2653752Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2654184Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2654574Z Traceback (most recent call last): 2025-12-04T09:36:41.2655052Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2655531Z check_stats(expected) 2025-12-04T09:36:41.2655929Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2656369Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2656990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2657683Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2658068Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2658281Z 2025-12-04T09:36:41.2658375Z Expected 196608 but got 0. 2025-12-04T09:36:41.2658638Z Absolute difference: 196608 2025-12-04T09:36:41.2658901Z Relative difference: 1.0 2025-12-04T09:36:41.2659069Z 2025-12-04T09:36:41.2659246Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2659925Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2660427Z 2025-12-04T09:36:41.2660747Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2661259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2661749Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2662170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2662603Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2663020Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2663454Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2663878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2664301Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2664727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2665149Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2665570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2666010Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2666433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2666869Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2667282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2667709Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2668121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2668544Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2668966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2669401Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2669824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2670248Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2670751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2671180Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2671596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2672016Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2672450Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2672843Z Traceback (most recent call last): 2025-12-04T09:36:41.2673308Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2673788Z check_stats(expected) 2025-12-04T09:36:41.2674184Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2674640Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2675246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2675952Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2676335Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2676544Z 2025-12-04T09:36:41.2676641Z Expected 196608 but got 0. 2025-12-04T09:36:41.2676900Z Absolute difference: 196608 2025-12-04T09:36:41.2677163Z Relative difference: 1.0 2025-12-04T09:36:41.2677327Z 2025-12-04T09:36:41.2677510Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2678192Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2678700Z 2025-12-04T09:36:41.2678923Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2679539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2679971Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2680399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2680857Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2681293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2681716Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2682134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2682561Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2682977Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2683404Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2683816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2684255Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2684666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2685105Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2685518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2685937Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2686114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2686261Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2686438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2686586Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2686760Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2686913Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2687084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2687232Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2687543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2687695Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2687866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2688012Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2688193Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2688303Z Traceback (most recent call last): 2025-12-04T09:36:41.2688587Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2688682Z check_stats(expected) 2025-12-04T09:36:41.2688929Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2689031Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2689459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2689632Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2689745Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2689749Z 2025-12-04T09:36:41.2689843Z Expected 196608 but got 0. 2025-12-04T09:36:41.2689938Z Absolute difference: 196608 2025-12-04T09:36:41.2690030Z Relative difference: 1.0 2025-12-04T09:36:41.2690035Z 2025-12-04T09:36:41.2690217Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2690617Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2690621Z 2025-12-04T09:36:41.2690844Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2691099Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2691271Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2691479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2691628Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2691799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2691954Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2692128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2692280Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2692451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2692595Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2692781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2692927Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2693102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2693252Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2693422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2693571Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2693740Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2693886Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2694061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2694206Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2694380Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2694535Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2694704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2694942Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2695116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2695262Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2695438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2695586Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2695757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2695907Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2696088Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2696207Z Traceback (most recent call last): 2025-12-04T09:36:41.2696488Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2696579Z check_stats(expected) 2025-12-04T09:36:41.2696826Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2696930Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2697349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2697525Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2697637Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2697641Z 2025-12-04T09:36:41.2697736Z Expected 196608 but got 0. 2025-12-04T09:36:41.2697831Z Absolute difference: 196608 2025-12-04T09:36:41.2697924Z Relative difference: 1.0 2025-12-04T09:36:41.2697928Z 2025-12-04T09:36:41.2698114Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2698595Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2698599Z 2025-12-04T09:36:41.2698827Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2699002Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2699153Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2699329Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2699476Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2699646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2699800Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2699971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2700123Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2700292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2700444Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2700620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2700767Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2700938Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2701088Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2701256Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2701412Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2701580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2701725Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2701904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2702051Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2702301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2702454Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2702625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2702774Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2702942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2703088Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2703260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2703404Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2703574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2703729Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2703902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2704053Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2704234Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2704340Z Traceback (most recent call last): 2025-12-04T09:36:41.2704628Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2704721Z check_stats(expected) 2025-12-04T09:36:41.2704956Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2705062Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2705481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2705743Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2705856Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2705860Z 2025-12-04T09:36:41.2705952Z Expected 196608 but got 0. 2025-12-04T09:36:41.2706060Z Absolute difference: 196608 2025-12-04T09:36:41.2706152Z Relative difference: 1.0 2025-12-04T09:36:41.2706156Z 2025-12-04T09:36:41.2706337Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2706747Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2706751Z 2025-12-04T09:36:41.2706971Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2707148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2707299Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2707469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2707631Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2707802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2707953Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2708128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2708274Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2708452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2708597Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2708766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2708920Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2709089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2709240Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2709419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2709756Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2709932Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2710078Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2710246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2710393Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2710563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2710712Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2710887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2711032Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2711209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2711360Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2711535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2711683Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2711852Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2712002Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2712176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2712321Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2712496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2712641Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2712905Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2713013Z Traceback (most recent call last): 2025-12-04T09:36:41.2713300Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2713391Z check_stats(expected) 2025-12-04T09:36:41.2713632Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2713733Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2714163Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2714335Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2714448Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2714453Z 2025-12-04T09:36:41.2714547Z Expected 196608 but got 0. 2025-12-04T09:36:41.2714641Z Absolute difference: 196608 2025-12-04T09:36:41.2714734Z Relative difference: 1.0 2025-12-04T09:36:41.2714750Z 2025-12-04T09:36:41.2714926Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2715329Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2715333Z 2025-12-04T09:36:41.2715554Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2715726Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2715874Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2716051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2716201Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2716376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2716522Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2716697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2716844Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2717097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2717246Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2717418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2717564Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2717738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2717883Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2718052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2718201Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2718369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2718518Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2718690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2718839Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2719014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2719159Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2719326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2719474Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2719646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2719791Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2719966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2720196Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2720369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2720515Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2720718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2720889Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2721062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2721212Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2721384Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2721529Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2721722Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2721829Z Traceback (most recent call last): 2025-12-04T09:36:41.2722114Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2722214Z check_stats(expected) 2025-12-04T09:36:41.2722458Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2722562Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2722986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2723162Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2723283Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2723288Z 2025-12-04T09:36:41.2723381Z Expected 196608 but got 0. 2025-12-04T09:36:41.2723479Z Absolute difference: 196608 2025-12-04T09:36:41.2723582Z Relative difference: 1.0 2025-12-04T09:36:41.2723586Z 2025-12-04T09:36:41.2723764Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2724165Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2724180Z 2025-12-04T09:36:41.2724400Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2724655Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2724809Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2724984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2725135Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2725310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2725455Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2725630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2725777Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2725951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2726102Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2726277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2726424Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2726600Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2726749Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2726927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2727073Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2727242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2727446Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2727619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2727851Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2728027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2728183Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2728355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2728507Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2728680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2728831Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2729003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2729147Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2729321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2729473Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2729643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2729797Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2729968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2730116Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2730287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2730432Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2730610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2730756Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2730941Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2731050Z Traceback (most recent call last): 2025-12-04T09:36:41.2731361Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2731466Z check_stats(expected) 2025-12-04T09:36:41.2731796Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2731902Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2732326Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2732501Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2732615Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2732625Z 2025-12-04T09:36:41.2732714Z Expected 196608 but got 0. 2025-12-04T09:36:41.2732807Z Absolute difference: 196608 2025-12-04T09:36:41.2732901Z Relative difference: 1.0 2025-12-04T09:36:41.2732906Z 2025-12-04T09:36:41.2733082Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2733487Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2733491Z 2025-12-04T09:36:41.2733710Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2733888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2734042Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2734214Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2734362Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2734537Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2734684Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2734854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2735004Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2735391Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2735543Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2735719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2735864Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2736039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2736183Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2736353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2736501Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2736670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2736819Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2736988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2737141Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2737313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2737462Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2737629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2737778Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2737955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2742872Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2743067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2743217Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2743395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2743552Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2743723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2744052Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2744225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2744376Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2744547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2744692Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2744862Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2745006Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2745177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2745325Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2745510Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2745621Z Traceback (most recent call last): 2025-12-04T09:36:41.2745909Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2746002Z check_stats(expected) 2025-12-04T09:36:41.2746246Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2746347Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2746769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2746945Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2747058Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2747063Z 2025-12-04T09:36:41.2747157Z Expected 196608 but got 0. 2025-12-04T09:36:41.2747250Z Absolute difference: 196608 2025-12-04T09:36:41.2747453Z Relative difference: 1.0 2025-12-04T09:36:41.2747458Z 2025-12-04T09:36:41.2747638Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2748042Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2748047Z 2025-12-04T09:36:41.2748266Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2748443Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2748592Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2748768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2748912Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2749082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2749230Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2749403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2749550Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2749727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2749873Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2750043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2750189Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2750355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2750506Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2750672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2750816Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2750991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2751139Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2751422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2751568Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2751735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2751886Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2752051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2752195Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2752368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2752512Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2752683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2752833Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2753001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2753151Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2753319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2753461Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2753633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2753777Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2753947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2754090Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2754258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2754488Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2754654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2754798Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2754976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2755119Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2755303Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2755409Z Traceback (most recent call last): 2025-12-04T09:36:41.2755692Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2755785Z check_stats(expected) 2025-12-04T09:36:41.2756018Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2756118Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2756540Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2756714Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2756832Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2756837Z 2025-12-04T09:36:41.2756927Z Expected 196608 but got 0. 2025-12-04T09:36:41.2757021Z Absolute difference: 196608 2025-12-04T09:36:41.2757116Z Relative difference: 1.0 2025-12-04T09:36:41.2757120Z 2025-12-04T09:36:41.2757295Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2757692Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2757700Z 2025-12-04T09:36:41.2757919Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2758088Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2758243Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2758411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2758556Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2758805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2758951Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2759121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2759265Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2759435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2759581Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2759748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2759894Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2760074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2760220Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2760394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2760539Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2760707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2760864Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2761057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2761218Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2761388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2761531Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2761703Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2761925Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2762092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2762246Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2762414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2762559Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2762737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2762879Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2763049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2763195Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2763363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2763517Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2763684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2763831Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2764003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2764148Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2764319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2764462Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2764635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2764783Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2764955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2765101Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2765292Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2765398Z Traceback (most recent call last): 2025-12-04T09:36:41.2765759Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2765850Z check_stats(expected) 2025-12-04T09:36:41.2766083Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2766189Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2766605Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2766776Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2766891Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2766896Z 2025-12-04T09:36:41.2766983Z Expected 196608 but got 0. 2025-12-04T09:36:41.2767079Z Absolute difference: 196608 2025-12-04T09:36:41.2767175Z Relative difference: 1.0 2025-12-04T09:36:41.2767180Z 2025-12-04T09:36:41.2767419Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2767827Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2767832Z 2025-12-04T09:36:41.2768046Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2768219Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2768369Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2768537Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2768688Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2768853Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2768997Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2769287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2769431Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2769604Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2769750Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2769919Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2770066Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2770234Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2770378Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2770549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2770718Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2770908Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2771062Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2771232Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2771379Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2771546Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2771691Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2771862Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2772006Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2772173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2772320Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2772487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2772644Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2772812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2773037Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2773210Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2773356Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2773524Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2773675Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2773842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2773989Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2774155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2774305Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2774477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2774620Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2774792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2774940Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2775106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2775252Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2775419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2775563Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2775748Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2775854Z Traceback (most recent call last): 2025-12-04T09:36:41.2776211Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2776307Z check_stats(expected) 2025-12-04T09:36:41.2776542Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2776655Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2777072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2777242Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2777356Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2777360Z 2025-12-04T09:36:41.2777449Z Expected 196608 but got 0. 2025-12-04T09:36:41.2777543Z Absolute difference: 196608 2025-12-04T09:36:41.2777637Z Relative difference: 1.0 2025-12-04T09:36:41.2777641Z 2025-12-04T09:36:41.2777818Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2778216Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2778226Z 2025-12-04T09:36:41.2778446Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2778619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2778772Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2778941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2779093Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2779261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2779406Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2779578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2779722Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2779896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2780043Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2780297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2780456Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2780649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2780816Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2780988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2781132Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2781299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2781446Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2781612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2781763Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2781931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2782081Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2782257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2782399Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2782567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2782714Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2782882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2783036Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2783205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2783429Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2783598Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2783741Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2783911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2784056Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2784224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2784372Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2784537Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2784683Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2784856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2784998Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2785171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2785320Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2785489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2785637Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2785805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2785950Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2786119Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2786262Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2786443Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2786551Z Traceback (most recent call last): 2025-12-04T09:36:41.2786827Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2786924Z check_stats(expected) 2025-12-04T09:36:41.2787155Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2787337Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2787762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2787931Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2788044Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2788048Z 2025-12-04T09:36:41.2788143Z Expected 196608 but got 0. 2025-12-04T09:36:41.2788234Z Absolute difference: 196608 2025-12-04T09:36:41.2788328Z Relative difference: 1.0 2025-12-04T09:36:41.2788332Z 2025-12-04T09:36:41.2788509Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2788903Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2788913Z 2025-12-04T09:36:41.2789132Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2789307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2789453Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2789626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2789770Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2789943Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2790089Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2790256Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2790405Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2790571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2790799Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2790995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2791168Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2791339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2791481Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2791648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2791795Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2791965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2792109Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2792279Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2792427Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2792599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2792747Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2792915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2793062Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2793228Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2793371Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2793541Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2793683Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2793853Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2793996Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2794168Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2794317Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2794565Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2794710Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2794879Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2795022Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2795191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2795336Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2795502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2795650Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2795824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2795968Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2796143Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2796286Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2796459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2796604Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2796773Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2796925Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2797096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2797240Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2797427Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2797637Z Traceback (most recent call last): 2025-12-04T09:36:41.2797919Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2798016Z check_stats(expected) 2025-12-04T09:36:41.2798252Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2798355Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2798772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2798942Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2799056Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2799060Z 2025-12-04T09:36:41.2799150Z Expected 196608 but got 0. 2025-12-04T09:36:41.2799249Z Absolute difference: 196608 2025-12-04T09:36:41.2799339Z Relative difference: 1.0 2025-12-04T09:36:41.2799343Z 2025-12-04T09:36:41.2799526Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2799926Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2799930Z 2025-12-04T09:36:41.2800148Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2800321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2800467Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2800634Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2800787Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2800955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2801106Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2801309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2801475Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2801646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2801872Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2802045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2802192Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2802361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2802507Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2802679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2802823Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2802994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2803145Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2803315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2803465Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2803640Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2803784Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2803958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2804102Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2804269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2804417Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2804588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2804736Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2804988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2805134Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2805309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2805455Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2805622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2805768Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2805935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2806087Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2806254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2806397Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2806570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2806721Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2806897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2807043Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2807211Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2807418Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2807586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2807729Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2807902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2808045Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2808215Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2808369Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2808536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2808771Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2808956Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2809060Z Traceback (most recent call last): 2025-12-04T09:36:41.2809340Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2809432Z check_stats(expected) 2025-12-04T09:36:41.2809669Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2809772Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2810188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2810362Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2810483Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2810487Z 2025-12-04T09:36:41.2810577Z Expected 196608 but got 0. 2025-12-04T09:36:41.2810672Z Absolute difference: 196608 2025-12-04T09:36:41.2810769Z Relative difference: 1.0 2025-12-04T09:36:41.2810773Z 2025-12-04T09:36:41.2810951Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2811348Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2811353Z 2025-12-04T09:36:41.2811568Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2811745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2811893Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2812067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2812296Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2812464Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2812616Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2812786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2812930Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2813104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2813250Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2813420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2813565Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2813733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2813883Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2814061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2814205Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2814380Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2814526Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2814697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2814842Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2815011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2815157Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2815326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2815471Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2815650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2815795Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2816048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2816195Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2816363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2816511Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2816679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2816825Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2816999Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2817143Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2817310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2817465Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2817637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2817789Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2817957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2818100Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2818274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2818418Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2818586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2818735Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2818904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2819130Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2819300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2819443Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2819622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2819766Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2819933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2820083Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2820251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2820395Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2820577Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2820680Z Traceback (most recent call last): 2025-12-04T09:36:41.2820998Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2821102Z check_stats(expected) 2025-12-04T09:36:41.2821347Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2821457Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2821873Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2822044Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2822154Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2822159Z 2025-12-04T09:36:41.2822246Z Expected 196608 but got 0. 2025-12-04T09:36:41.2822343Z Absolute difference: 196608 2025-12-04T09:36:41.2822434Z Relative difference: 1.0 2025-12-04T09:36:41.2822438Z 2025-12-04T09:36:41.2822613Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2823014Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2823023Z 2025-12-04T09:36:41.2823238Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2823521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2823672Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2823842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2823990Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2824159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2824310Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2824479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2824624Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2824800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2824945Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2825123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2825271Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2825439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2825585Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2825752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2825895Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2826064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2826207Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2826375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2826602Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2826770Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2826923Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2827088Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2827232Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2827411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2827555Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2827721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2827868Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2828036Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2828188Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2828354Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2828502Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2828673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2828816Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2828982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2829132Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2829298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2829445Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2829612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2829756Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2829931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2830074Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2830403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2830553Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2830729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2830897Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2831088Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2831231Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2831401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2831544Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2831715Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2831863Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2832034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2832180Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2832348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2832494Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2832677Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2832781Z Traceback (most recent call last): 2025-12-04T09:36:41.2833059Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2833152Z check_stats(expected) 2025-12-04T09:36:41.2833384Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2833568Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2833986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2834165Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2834277Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2834282Z 2025-12-04T09:36:41.2834370Z Expected 196608 but got 0. 2025-12-04T09:36:41.2834468Z Absolute difference: 196608 2025-12-04T09:36:41.2834563Z Relative difference: 1.0 2025-12-04T09:36:41.2834567Z 2025-12-04T09:36:41.2834743Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2835143Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2835147Z 2025-12-04T09:36:41.2835598Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2835772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2835927Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2836095Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2836255Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2836423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2836568Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2836737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2836879Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2837048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2837195Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2837362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2837518Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2837685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2837950Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2838124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2838268Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2838437Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2838582Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2838754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2838901Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2839071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2839214Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2839392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2839539Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2839716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2839865Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2840032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2840176Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2840347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2840490Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2840660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2840804Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2841113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2841279Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2841452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2841597Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2841766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2841909Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2842077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2842220Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2842387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2842532Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2842698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2842846Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2843015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2843161Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2843332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2843475Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2843641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2843788Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2843956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2844099Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2844268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2844417Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2844586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2844812Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2844981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2845127Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2845309Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2845414Z Traceback (most recent call last): 2025-12-04T09:36:41.2845692Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2845783Z check_stats(expected) 2025-12-04T09:36:41.2846021Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2846127Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2846543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2846722Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2846837Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2846842Z 2025-12-04T09:36:41.2846930Z Expected 196608 but got 0. 2025-12-04T09:36:41.2847025Z Absolute difference: 196608 2025-12-04T09:36:41.2847114Z Relative difference: 1.0 2025-12-04T09:36:41.2847118Z 2025-12-04T09:36:41.2847338Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2847735Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2847739Z 2025-12-04T09:36:41.2847953Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2848126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2848356Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2848531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2848677Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2848849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2848996Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2849162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2849306Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2849475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2849619Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2849789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2849934Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2850106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2850253Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2850422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2850565Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2850734Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2850878Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2851052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2851195Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2851360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2851506Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2851672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2851819Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2851990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2852239Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2852415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2852561Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2852729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2852878Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2853045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2853187Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2853358Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2853505Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2853675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2853824Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2853991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2854140Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2854306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2854450Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2854620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2854765Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2854939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2855082Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2855330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2855478Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2855650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2855795Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2855965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2856108Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2856276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2856419Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2856587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2856734Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2856906Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2857051Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2857225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2857372Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2857545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2857688Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2857869Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2857977Z Traceback (most recent call last): 2025-12-04T09:36:41.2858252Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2858341Z check_stats(expected) 2025-12-04T09:36:41.2858577Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2858685Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2859104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2859356Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2859470Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2859474Z 2025-12-04T09:36:41.2859567Z Expected 196608 but got 0. 2025-12-04T09:36:41.2859660Z Absolute difference: 196608 2025-12-04T09:36:41.2859750Z Relative difference: 1.0 2025-12-04T09:36:41.2859755Z 2025-12-04T09:36:41.2859933Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2860330Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2860334Z 2025-12-04T09:36:41.2860578Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2860780Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2860926Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2861104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2861249Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2861416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2861564Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2861731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2861877Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2862044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2862188Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2862357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2862582Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2862750Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2862903Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2863070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2863216Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2863382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2863525Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2863696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2863844Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2864012Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2864163Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2864329Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2864475Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2864648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2864791Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2864961Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2865105Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2865272Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2865419Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2865584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2865735Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2865913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2866056Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2866318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2866462Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2866630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2866781Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2866948Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2867093Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2867258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2867401Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2867569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2867719Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2867886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2868038Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2868204Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2868349Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2868515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2868658Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2868831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2873500Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2873695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2873960Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2874135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2874291Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2874459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2874604Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2874780Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2874924Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2875093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2875244Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2875424Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2875532Z Traceback (most recent call last): 2025-12-04T09:36:41.2875817Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2875909Z check_stats(expected) 2025-12-04T09:36:41.2876147Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2876247Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2876671Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2876848Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2876961Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2876967Z 2025-12-04T09:36:41.2877061Z Expected 196608 but got 0. 2025-12-04T09:36:41.2877156Z Absolute difference: 196608 2025-12-04T09:36:41.2877247Z Relative difference: 1.0 2025-12-04T09:36:41.2877251Z 2025-12-04T09:36:41.2877432Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2877840Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2877845Z 2025-12-04T09:36:41.2878149Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2878322Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2878468Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2878639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2878783Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2878951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2879098Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2879265Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2879415Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2879589Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2879733Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2879909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2880055Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2880222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2880370Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2880536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2880683Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2880849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2880991Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2881240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2881383Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2881555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2881706Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2881873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2882019Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2882185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2882332Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2882502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2882645Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2882811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2882963Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2883129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2883280Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2883446Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2883589Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2883759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2883902Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2884068Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2884214Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2884379Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2884531Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2884697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2884841Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2885117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2885262Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2885428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2885574Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2885739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2885885Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2886052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2886195Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2886369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2886515Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2886689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2886834Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2887000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2887145Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2887454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2887598Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2887766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2887911Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2888077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2888312Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2888479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2888631Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2888811Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2888914Z Traceback (most recent call last): 2025-12-04T09:36:41.2889196Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2889285Z check_stats(expected) 2025-12-04T09:36:41.2889517Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2889622Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2890040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2890217Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2890329Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2890334Z 2025-12-04T09:36:41.2890422Z Expected 196608 but got 0. 2025-12-04T09:36:41.2890525Z Absolute difference: 196608 2025-12-04T09:36:41.2890616Z Relative difference: 1.0 2025-12-04T09:36:41.2890620Z 2025-12-04T09:36:41.2890794Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2891205Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2891211Z 2025-12-04T09:36:41.2891468Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2891641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2891786Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2891955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2892107Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2892276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2892504Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2892675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2892818Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2892988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2893132Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2893300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2893444Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2893610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2893757Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2893930Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2894077Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2894246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2894389Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2894557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2894702Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2894868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2895010Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2895181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2895323Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2895574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2895719Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2895890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2896043Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2896212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2896355Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2896524Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2896668Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2896836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2896980Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2897158Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2897303Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2897473Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2897617Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2897786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2897933Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2898102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2898244Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2898410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2898554Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2898720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2898870Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2899039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2899264Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2899437Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2899579Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2899745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2899891Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2900056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2900199Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2900369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2900521Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2900690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2900837Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2901039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2901201Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2901367Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2901511Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2901680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2901823Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2901992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2902134Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2902400Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2902506Z Traceback (most recent call last): 2025-12-04T09:36:41.2902788Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2902878Z check_stats(expected) 2025-12-04T09:36:41.2903114Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2903214Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2903631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2903799Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2903910Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2903914Z 2025-12-04T09:36:41.2904005Z Expected 196608 but got 0. 2025-12-04T09:36:41.2904101Z Absolute difference: 196608 2025-12-04T09:36:41.2904190Z Relative difference: 1.0 2025-12-04T09:36:41.2904200Z 2025-12-04T09:36:41.2904376Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2904777Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2904781Z 2025-12-04T09:36:41.2905000Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2905170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2905315Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2905485Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2905634Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2905804Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2905949Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2906126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2906273Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2906519Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2906663Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2906835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2906978Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2907147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2907289Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2907454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2907600Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2907771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2907919Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2908092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2908244Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2908414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2908556Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2908722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2908866Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2909032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2909174Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2909343Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2909567Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2909736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2909884Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2910051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2910202Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2910370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2910518Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2910688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2910831Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2911001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2911144Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2911321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2911467Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2911644Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2911786Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2911956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2912103Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2912277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2912420Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2912586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2912733Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2912905Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2913052Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2913328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2913473Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2913642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2913784Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2913950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2914095Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2914262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2914404Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2914576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2914721Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2914889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2915036Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2915208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2915352Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2915518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2915665Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2915837Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2915979Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2916158Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2916346Z Traceback (most recent call last): 2025-12-04T09:36:41.2916622Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2916713Z check_stats(expected) 2025-12-04T09:36:41.2916950Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2917050Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2917467Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2917634Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2917744Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2917752Z 2025-12-04T09:36:41.2917839Z Expected 196608 but got 0. 2025-12-04T09:36:41.2917930Z Absolute difference: 196608 2025-12-04T09:36:41.2918022Z Relative difference: 1.0 2025-12-04T09:36:41.2918027Z 2025-12-04T09:36:41.2918200Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2918601Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2918606Z 2025-12-04T09:36:41.2918826Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2918995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2919143Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2919310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2919455Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2919624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2919766Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2919932Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2920079Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2920249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2920395Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2920642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2920805Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2921000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2921147Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2921313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2921458Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2921624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2921769Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2921942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2922084Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2922258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2922401Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2922567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2922714Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2922880Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2923025Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2923191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2923333Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2923501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2923804Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2923974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2924124Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2924296Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2924445Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2924611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2924753Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2924921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2925063Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2925229Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2925380Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2925545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2925699Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2925864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2926007Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2926175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2926317Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2926483Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2926629Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2926795Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2926944Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2927115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2927258Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2927557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2927701Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2927866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2928014Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2928181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2928326Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2928491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2928636Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2928806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2928959Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2929131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2929279Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2929445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2929591Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2929758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2929900Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2930069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2930212Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2930392Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2930581Z Traceback (most recent call last): 2025-12-04T09:36:41.2930858Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2930952Z check_stats(expected) 2025-12-04T09:36:41.2931221Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2931342Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2931760Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2931929Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2932039Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2932044Z 2025-12-04T09:36:41.2932135Z Expected 196608 but got 0. 2025-12-04T09:36:41.2932227Z Absolute difference: 196608 2025-12-04T09:36:41.2932320Z Relative difference: 1.0 2025-12-04T09:36:41.2932325Z 2025-12-04T09:36:41.2932499Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2932896Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2932902Z 2025-12-04T09:36:41.2933125Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2933298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2933445Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2933614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2933759Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2933928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2934071Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2934238Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2934389Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2934556Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2934782Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2934953Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2935097Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2935441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2935590Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2935756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2935901Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2936072Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2936218Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2936390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2936532Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2936709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2936852Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2937022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2937167Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2937333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2937478Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2937644Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2937790Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2938120Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2938263Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2938437Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2938583Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2938748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2938897Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2939063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2939205Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2939374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2939517Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2939683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2939837Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2940004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2940159Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2940327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2940472Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2940641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2940784Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2940950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2941109Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2941303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2941462Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2941629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2941771Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2942051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2942196Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2942361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2942508Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2942676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2942822Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2942988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2943130Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2943307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2943449Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2943623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2943768Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2943937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2944084Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2944251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2944395Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2944563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2944706Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2944871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2945098Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2945283Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2945387Z Traceback (most recent call last): 2025-12-04T09:36:41.2945666Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2945755Z check_stats(expected) 2025-12-04T09:36:41.2945988Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2946089Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2946505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2946676Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2946786Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2946797Z 2025-12-04T09:36:41.2946889Z Expected 196608 but got 0. 2025-12-04T09:36:41.2946981Z Absolute difference: 196608 2025-12-04T09:36:41.2947072Z Relative difference: 1.0 2025-12-04T09:36:41.2947077Z 2025-12-04T09:36:41.2947259Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2947657Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2947661Z 2025-12-04T09:36:41.2947875Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2948046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2948191Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2948361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2948505Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2948671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2948823Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2948989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2949212Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2949384Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2949527Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2949698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2949840Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2950005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2950151Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2950316Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2950463Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2950632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2950778Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2950947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2951090Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2951259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2951406Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2951572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2951715Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2951883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2952026Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2952274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2952417Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2952588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2952735Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2952901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2953044Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2953212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2953354Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2953522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2953666Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2953841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2953986Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2954155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2954298Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2954468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2954611Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2954781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2954922Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2955089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2955235Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2955401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2955549Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2955718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2955941Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2956111Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2956253Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2956419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2956564Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2956730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2956877Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2957046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2957193Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2957362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2957504Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2957677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2957822Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2957987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2958131Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2958300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2958446Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2958615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2958758Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2959006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2959151Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2959325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2959468Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2959638Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2959780Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2959963Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2960069Z Traceback (most recent call last): 2025-12-04T09:36:41.2960345Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2960438Z check_stats(expected) 2025-12-04T09:36:41.2960669Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2960777Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2961219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2961416Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2961530Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2961534Z 2025-12-04T09:36:41.2961622Z Expected 196608 but got 0. 2025-12-04T09:36:41.2961714Z Absolute difference: 196608 2025-12-04T09:36:41.2961810Z Relative difference: 1.0 2025-12-04T09:36:41.2961815Z 2025-12-04T09:36:41.2961988Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2962383Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2962390Z 2025-12-04T09:36:41.2962606Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2962781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2962929Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2963177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2963324Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2963494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2963638Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2963808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2963952Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2964119Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2964267Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2964433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2964583Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2964752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2964904Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2965073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2965216Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2965381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2965526Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2965692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2965834Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2966003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2966250Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2966419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2966567Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2966732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2966879Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2967045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2967188Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2967407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2967550Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2967715Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2967866Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2968042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2968186Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2968356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2968498Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2968668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2968813Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2968979Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2969125Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2969293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2969439Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2969610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2969753Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2970007Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2970153Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2970319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2970465Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2970631Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2970776Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2970943Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2971085Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2971254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2971403Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2971569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2971719Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2971886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2972030Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2972196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2972339Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2972508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2972650Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2972817Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2973040Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2973207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2973353Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2973527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2973669Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2973838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2973980Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2974147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2974296Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2974462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2974608Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2974779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2974922Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2975112Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2975216Z Traceback (most recent call last): 2025-12-04T09:36:41.2975493Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2975585Z check_stats(expected) 2025-12-04T09:36:41.2975816Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2975918Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2976333Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2976501Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2976615Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2976624Z 2025-12-04T09:36:41.2976712Z Expected 196608 but got 0. 2025-12-04T09:36:41.2976805Z Absolute difference: 196608 2025-12-04T09:36:41.2976902Z Relative difference: 1.0 2025-12-04T09:36:41.2976906Z 2025-12-04T09:36:41.2977164Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2977564Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2977569Z 2025-12-04T09:36:41.2977783Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2977952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2978100Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2978268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2978416Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2978586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2978730Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2978907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2979051Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2979216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2979361Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2979527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2979672Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2979837Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2979981Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2980149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2980374Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2980540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2980694Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2980860Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2981006Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2981200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2981367Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2981538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2981681Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2981846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2981999Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2982164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2982315Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2982481Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2982624Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2982792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2982938Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2983103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2983249Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2983415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2983565Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2983737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2983881Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2984145Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2984290Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2984456Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2984602Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2984768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2984915Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2985082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2985225Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2985405Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2985548Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2985722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2985868Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2986034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2986178Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2986345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2986488Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2986657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2986800Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2986970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2987199Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2987366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2987517Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2987684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2987826Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2988000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2988143Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2988313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2988460Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2988625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2988777Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2988944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2989086Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2989259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2989402Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2989570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2989717Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2989882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2990029Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2990196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2990338Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2990525Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.2990628Z Traceback (most recent call last): 2025-12-04T09:36:41.2991007Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.2991101Z check_stats(expected) 2025-12-04T09:36:41.2991332Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.2991435Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.2991849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.2992017Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.2992129Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.2992133Z 2025-12-04T09:36:41.2992223Z Expected 196608 but got 0. 2025-12-04T09:36:41.2992314Z Absolute difference: 196608 2025-12-04T09:36:41.2992412Z Relative difference: 1.0 2025-12-04T09:36:41.2992417Z 2025-12-04T09:36:41.2992591Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.2992991Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.2992996Z 2025-12-04T09:36:41.2993211Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.2993380Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2993528Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2993695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2993839Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2994009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2994152Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2994399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2994543Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2994716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2994862Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2995029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2995171Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2995344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2995491Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2995659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2995802Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2995968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2996120Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2996287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2996437Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2996607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2996753Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2996922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2997064Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2997230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2997376Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.2997544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.2997698Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3001280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3001538Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3001717Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3001866Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3002037Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3002185Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3002356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3002500Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3002673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3002817Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3002997Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3003141Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3003317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3003464Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3003632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3003776Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3003946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3004091Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3004267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3004414Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3004662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3004809Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3004984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3005128Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3005299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3005445Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3005616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3005759Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3005928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3006075Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3006243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3006395Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3006568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3006715Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3006889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3007033Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3007202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3007416Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3007585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3007729Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3007901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3008051Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3008222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3008367Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3008618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3008767Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3008936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3009081Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3009253Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3009398Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3009569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3009714Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3009901Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3010010Z Traceback (most recent call last): 2025-12-04T09:36:41.3010299Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3010391Z check_stats(expected) 2025-12-04T09:36:41.3010629Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3010731Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3011205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3011378Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3011493Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3011498Z 2025-12-04T09:36:41.3011592Z Expected 196608 but got 0. 2025-12-04T09:36:41.3011686Z Absolute difference: 196608 2025-12-04T09:36:41.3011856Z Relative difference: 1.0 2025-12-04T09:36:41.3011861Z 2025-12-04T09:36:41.3012041Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3012447Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3012452Z 2025-12-04T09:36:41.3012674Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3012848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3012995Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3013166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3013311Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3013481Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3013625Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3013800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3013948Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3014119Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3014262Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3014432Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3014575Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3014743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3014885Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3015051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3015195Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3015362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3015511Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3015681Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3015908Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3016078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3016221Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3016387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3016532Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3016699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3016841Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3017010Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3017157Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3017330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3017479Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3017645Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3017796Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3017963Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3018106Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3018274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3018422Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3018593Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3018735Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3019039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3019188Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3019364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3019506Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3019679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3019821Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3019990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3020133Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3020301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3020447Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3020619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3020761Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3020933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3021075Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3021244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3021386Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3021554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3021700Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3021866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3022008Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3022177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3022327Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3022494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3022746Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3022917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3023062Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3023230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3023375Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3023549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3023692Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3023857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3024011Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3024177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3024328Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3024500Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3024643Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3024812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3024955Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3025121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3025268Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3025435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3025580Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3025827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3025969Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3026156Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3026261Z Traceback (most recent call last): 2025-12-04T09:36:41.3026539Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3026633Z check_stats(expected) 2025-12-04T09:36:41.3026864Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3026967Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3027383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3027552Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3027664Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3027675Z 2025-12-04T09:36:41.3027764Z Expected 196608 but got 0. 2025-12-04T09:36:41.3027857Z Absolute difference: 196608 2025-12-04T09:36:41.3027951Z Relative difference: 1.0 2025-12-04T09:36:41.3027955Z 2025-12-04T09:36:41.3028136Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3028535Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3028540Z 2025-12-04T09:36:41.3028754Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3028923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3029071Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3029238Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3029385Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3029556Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3029700Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3029946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3030091Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3030258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3030405Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3030571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3030745Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3030936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3031078Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3031247Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3031397Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3031563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3031719Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3031885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3032035Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3032200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3032342Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3032511Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3032653Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3032819Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3033047Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3033216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3033373Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3033539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3033683Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3033851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3033993Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3034161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3034306Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3034473Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3034619Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3034790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3034935Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3035108Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3035504Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3035675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3035822Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3035988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3036133Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3036301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3036443Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3036618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3036760Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3037134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3037286Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3037452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3037598Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3037765Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3037912Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3038087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3038230Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3038395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3038546Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3038712Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3038863Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3039029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3039171Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3039339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3039482Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3039648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3039799Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3039965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3040217Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3040385Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3040530Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3040713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3040856Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3041022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3041168Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3041360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3041530Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3041697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3041839Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3042014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3042157Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3042325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3042470Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3042636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3042784Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3042964Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3043067Z Traceback (most recent call last): 2025-12-04T09:36:41.3043351Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3043440Z check_stats(expected) 2025-12-04T09:36:41.3043672Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3043779Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3044194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3044448Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3044560Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3044564Z 2025-12-04T09:36:41.3044652Z Expected 196608 but got 0. 2025-12-04T09:36:41.3044746Z Absolute difference: 196608 2025-12-04T09:36:41.3044837Z Relative difference: 1.0 2025-12-04T09:36:41.3044841Z 2025-12-04T09:36:41.3045016Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3045413Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3045417Z 2025-12-04T09:36:41.3045632Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3045807Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3045956Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3046130Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3046276Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3046443Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3046586Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3046755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3046897Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3047070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3047213Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3047422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3047676Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3047844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3047992Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3048163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3048306Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3048474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3048616Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3048781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3048926Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3049091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3049238Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3049407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3049553Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3049722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3049864Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3050030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3050180Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3050345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3050489Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3050657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3050800Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3050974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3051117Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3051364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3051510Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3051676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3051818Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3051989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3052131Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3052299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3052441Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3052612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3052758Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3052931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3053073Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3053243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3053389Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3053558Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3053702Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3053868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3054016Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3054182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3054408Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3054581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3054732Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3054906Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3055050Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3055217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3055364Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3055536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3055678Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3055848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3055997Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3056167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3056312Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3056482Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3056632Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3056799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3056943Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3057112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3057256Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3057425Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3057569Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3057742Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3057892Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3058140Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3058285Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3058457Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3058602Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3058774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3058919Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3059088Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3059236Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3059405Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3059556Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3059739Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3059852Z Traceback (most recent call last): 2025-12-04T09:36:41.3060134Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3060222Z check_stats(expected) 2025-12-04T09:36:41.3060459Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3060573Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3061036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3061206Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3061319Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3061324Z 2025-12-04T09:36:41.3061497Z Expected 196608 but got 0. 2025-12-04T09:36:41.3061591Z Absolute difference: 196608 2025-12-04T09:36:41.3061688Z Relative difference: 1.0 2025-12-04T09:36:41.3061692Z 2025-12-04T09:36:41.3061874Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3062273Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3062277Z 2025-12-04T09:36:41.3062493Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3062664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3062812Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3062983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3063132Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3063299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3063451Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3063625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3063775Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3063945Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3064088Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3064254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3064403Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3064569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3064713Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3064882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3065032Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3065202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3065424Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3065593Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3065739Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3065906Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3066050Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3066221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3066365Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3066533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3066679Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3066854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3067003Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3067176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3067321Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3067491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3067635Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3067806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3067950Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3068117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3068264Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3068512Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3068655Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3068831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3068974Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3069149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3069293Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3069459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3069605Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3069774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3069918Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3070090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3070239Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3070409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3070556Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3070723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3070868Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3071036Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3071179Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3071349Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3071493Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3071663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3071811Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3071979Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3072125Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3072373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3072519Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3072690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3072833Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3073002Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3073150Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3073316Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3073465Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3073636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3073779Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3073955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3074101Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3074272Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3074416Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3074584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3074730Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3074896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3075040Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3075209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3075461Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3075631Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3075783Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3075952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3076099Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3076269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3076412Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3076583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3076726Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3076909Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3077018Z Traceback (most recent call last): 2025-12-04T09:36:41.3077295Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3077386Z check_stats(expected) 2025-12-04T09:36:41.3077624Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3077724Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3078144Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3078314Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3078429Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3078433Z 2025-12-04T09:36:41.3078521Z Expected 196608 but got 0. 2025-12-04T09:36:41.3078614Z Absolute difference: 196608 2025-12-04T09:36:41.3078707Z Relative difference: 1.0 2025-12-04T09:36:41.3078712Z 2025-12-04T09:36:41.3078888Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3079293Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3079297Z 2025-12-04T09:36:41.3079598Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3079770Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3079919Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3080087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3080232Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3080402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3080546Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3080713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3080893Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3081085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3081242Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3081410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3081553Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3081723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3081867Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3082035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3082181Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3082348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3082495Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3082746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3082890Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3083065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3083209Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3083376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3083523Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3083690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3083835Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3084001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3084145Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3084321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3084464Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3084633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3084780Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3084947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3085093Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3085261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3085403Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3085573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3085716Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3085883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3086039Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3086207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3086436Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3086605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3086749Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3086919Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3087063Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3087230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3087424Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3087595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3087749Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3087917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3088060Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3088235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3088379Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3088546Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3088692Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3088859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3089005Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3089174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3089319Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3089570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3089714Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3089888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3090034Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3090201Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3090347Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3090514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3090657Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3090826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3090969Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3091136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3091291Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3091459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3091607Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3091775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3091918Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3092088Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3092231Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3092398Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3092544Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3092711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3092864Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3093032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3093255Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3093426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3093570Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3093737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3093886Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3094056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3094204Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3094384Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3094487Z Traceback (most recent call last): 2025-12-04T09:36:41.3094769Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3094860Z check_stats(expected) 2025-12-04T09:36:41.3095104Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3095207Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3095624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3095797Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3095907Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3095912Z 2025-12-04T09:36:41.3095999Z Expected 196608 but got 0. 2025-12-04T09:36:41.3096099Z Absolute difference: 196608 2025-12-04T09:36:41.3096189Z Relative difference: 1.0 2025-12-04T09:36:41.3096193Z 2025-12-04T09:36:41.3096369Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3096849Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3096853Z 2025-12-04T09:36:41.3097076Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3097247Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3097392Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3097561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3097708Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3097878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3098025Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3098193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3098337Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3098511Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3098655Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3098826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3098972Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3099139Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3099286Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3099454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3099598Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3099768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3099911Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3100079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3100230Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3100541Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3100713Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3100902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3101046Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3101217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3101361Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3101529Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3101677Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3101848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3101998Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3102165Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3102316Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3102487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3102633Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3102801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3102948Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3103115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3103261Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3103428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3103653Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3103824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3103968Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3104141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3104288Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3104455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3104602Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3104769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3104912Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3105083Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3105226Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3105400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3105547Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3105719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3105865Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3106033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3106177Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3106347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3106489Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3106656Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3106802Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3106969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3107123Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3107291Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3107512Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3107684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3107828Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3107995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3108141Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3108308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3108454Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3108623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3108772Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3108942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3109090Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3109257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3109404Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3109571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3109718Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3109886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3110029Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3110200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3110344Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3110591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3110738Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3110911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3111058Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3111225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3111368Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3111538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3111682Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3111849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3111995Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3112180Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3112284Z Traceback (most recent call last): 2025-12-04T09:36:41.3112571Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3112662Z check_stats(expected) 2025-12-04T09:36:41.3112899Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3112999Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3113417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3113591Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3113702Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3113707Z 2025-12-04T09:36:41.3113798Z Expected 196608 but got 0. 2025-12-04T09:36:41.3113891Z Absolute difference: 196608 2025-12-04T09:36:41.3113982Z Relative difference: 1.0 2025-12-04T09:36:41.3113993Z 2025-12-04T09:36:41.3114171Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3114649Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3114655Z 2025-12-04T09:36:41.3114872Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3115046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3115191Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3115363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3115508Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3115678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3115825Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3115994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3116146Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3116322Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3116466Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3116636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3116780Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3116948Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3117096Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3117263Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3117407Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3117581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3117806Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3117975Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3118125Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3118292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3118439Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3118607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3118751Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3118921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3119065Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3119235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3119378Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3119553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3119702Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3119873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3120017Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3120187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3120331Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3120500Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3120644Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3120811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3120957Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3121129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3121273Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3121522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3121668Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3121845Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3121989Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3122155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3122302Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3122469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3122613Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3122785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3122935Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3123104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3123251Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3123418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3123563Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3123731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3123876Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3124046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3124190Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3124360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3124585Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3124752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3124906Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3128421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3128590Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3128769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3128918Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3129095Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3129242Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3129411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3129561Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3129743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3129890Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3130069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3130214Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3130388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3130551Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3130747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3130901Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3131071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3131217Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3131395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3131541Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3131848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3131996Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3132166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3132314Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3132485Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3132630Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3132803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3132948Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3133120Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3133271Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3133441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3133598Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3133782Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3133891Z Traceback (most recent call last): 2025-12-04T09:36:41.3134179Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3134272Z check_stats(expected) 2025-12-04T09:36:41.3134513Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3134616Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3135040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3135217Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3135604Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3135610Z 2025-12-04T09:36:41.3135699Z Expected 196608 but got 0. 2025-12-04T09:36:41.3135801Z Absolute difference: 196608 2025-12-04T09:36:41.3135891Z Relative difference: 1.0 2025-12-04T09:36:41.3135895Z 2025-12-04T09:36:41.3136077Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3136475Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3136480Z 2025-12-04T09:36:41.3136698Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3136874Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3137024Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3137193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3137347Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3137517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3137667Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3137835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3137978Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3138149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3138293Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3138460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3138611Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3138778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3138924Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3139098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3139241Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3139536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3139683Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3139851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3139997Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3140170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3140315Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3140484Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3140627Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3140803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3140946Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3141118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3141267Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3141433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3141579Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3141746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3141890Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3142060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3142203Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3142370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3142615Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3142782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3142933Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3143106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3143248Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3143419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3143562Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3143729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3143876Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3144043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3144194Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3144361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3144508Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3144678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3144822Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3144989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3145138Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3145311Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3145458Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3145624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3145767Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3145944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3146089Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3146335Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3146482Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3146650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3146796Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3146963Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3147106Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3147276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3147419Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3147590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3147739Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3147917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3148063Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3148231Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3148373Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3148545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3148688Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3148854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3149001Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3149170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3149397Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3149563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3149709Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3149878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3150020Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3150185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3150331Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3150497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3150646Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3150813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3150960Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3151128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3151271Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3151442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3151588Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3151755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3151901Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3152065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3152210Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3152392Z _______________________ TestCuda.test_host_memory_stats ________________________ 2025-12-04T09:36:41.3152497Z Traceback (most recent call last): 2025-12-04T09:36:41.3152777Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 223, in test_host_memory_stats 2025-12-04T09:36:41.3152869Z check_stats(expected) 2025-12-04T09:36:41.3153257Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 209, in check_stats 2025-12-04T09:36:41.3153362Z self.assertEqual(v, stats[k]) 2025-12-04T09:36:41.3153778Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3153949Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3154060Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3154065Z 2025-12-04T09:36:41.3154153Z Expected 196608 but got 0. 2025-12-04T09:36:41.3154246Z Absolute difference: 196608 2025-12-04T09:36:41.3154339Z Relative difference: 1.0 2025-12-04T09:36:41.3154343Z 2025-12-04T09:36:41.3154516Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3154913Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3154922Z 2025-12-04T09:36:41.3155137Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3155311Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3155464Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3155632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3155778Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3155945Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3156088Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3156256Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3156399Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3156669Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3156816Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3156987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3157133Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3157299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3157441Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3157610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3157753Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3157920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3158066Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3158234Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3158384Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3158550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3158695Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3158864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3159007Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3159173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3159318Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3159488Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3159634Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3159800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3159947Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3160115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3160338Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3160506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3160662Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3160857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3161023Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3161190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3161332Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3161502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3161644Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3161815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3161961Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3162132Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3162277Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3162444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3162586Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3162754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3162899Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3163064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3163213Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3163458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3163603Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3163774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3163917Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3164086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3164232Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3164399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3164544Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3164710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3164859Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3165029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3165178Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3165347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3165498Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3165664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3165808Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3165973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3166119Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3166286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3166428Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3166596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3166745Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3166912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3167057Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3167355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3167502Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3167677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3167821Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3167989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3168131Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3168297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3168442Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3168614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3168757Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3168936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3169079Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3169246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3169388Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3169553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3169699Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3169869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3170012Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3170180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3170408Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3170577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3170728Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3170895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:36:41.3171059Z key: allocated_bytes.freed, expected: 0, stats: 196608 2025-12-04T09:36:41.3171297Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3171400Z Traceback (most recent call last): 2025-12-04T09:36:41.3171722Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3171856Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3172277Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3172449Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3172559Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3172563Z 2025-12-04T09:36:41.3172657Z Expected 0 but got 123. 2025-12-04T09:36:41.3172748Z Absolute difference: 123 2025-12-04T09:36:41.3172838Z Relative difference: inf 2025-12-04T09:36:41.3172845Z 2025-12-04T09:36:41.3173020Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3173466Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3173471Z 2025-12-04T09:36:41.3173690Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3173895Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3173997Z Traceback (most recent call last): 2025-12-04T09:36:41.3174323Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3174460Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3174957Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3175127Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3175237Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3175240Z 2025-12-04T09:36:41.3175331Z Expected 0 but got 123. 2025-12-04T09:36:41.3175421Z Absolute difference: 123 2025-12-04T09:36:41.3175511Z Relative difference: inf 2025-12-04T09:36:41.3175519Z 2025-12-04T09:36:41.3175694Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3176138Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3176150Z 2025-12-04T09:36:41.3176366Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3176572Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3176680Z Traceback (most recent call last): 2025-12-04T09:36:41.3177000Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3177131Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3177546Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3177714Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3177823Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3177827Z 2025-12-04T09:36:41.3177914Z Expected 0 but got 123. 2025-12-04T09:36:41.3178006Z Absolute difference: 123 2025-12-04T09:36:41.3178096Z Relative difference: inf 2025-12-04T09:36:41.3178180Z 2025-12-04T09:36:41.3178359Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3178811Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3178816Z 2025-12-04T09:36:41.3179032Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3179237Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3179339Z Traceback (most recent call last): 2025-12-04T09:36:41.3179660Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3179791Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3180205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3180372Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3180488Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3180492Z 2025-12-04T09:36:41.3180582Z Expected 0 but got 123. 2025-12-04T09:36:41.3180673Z Absolute difference: 123 2025-12-04T09:36:41.3180766Z Relative difference: inf 2025-12-04T09:36:41.3180770Z 2025-12-04T09:36:41.3180950Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3181389Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3181393Z 2025-12-04T09:36:41.3181610Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3181815Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3181916Z Traceback (most recent call last): 2025-12-04T09:36:41.3182244Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3182380Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3182796Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3183045Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3183155Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3183159Z 2025-12-04T09:36:41.3183254Z Expected 0 but got 123. 2025-12-04T09:36:41.3183344Z Absolute difference: 123 2025-12-04T09:36:41.3183434Z Relative difference: inf 2025-12-04T09:36:41.3183438Z 2025-12-04T09:36:41.3183617Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3184055Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3184059Z 2025-12-04T09:36:41.3184275Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3184484Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3184586Z Traceback (most recent call last): 2025-12-04T09:36:41.3184910Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3185041Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3185458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3185624Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3185734Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3185737Z 2025-12-04T09:36:41.3185828Z Expected 0 but got 123. 2025-12-04T09:36:41.3185919Z Absolute difference: 123 2025-12-04T09:36:41.3186009Z Relative difference: inf 2025-12-04T09:36:41.3186013Z 2025-12-04T09:36:41.3186191Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3186736Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3186740Z 2025-12-04T09:36:41.3186964Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3187170Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3187273Z Traceback (most recent call last): 2025-12-04T09:36:41.3187592Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3187722Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3188138Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3188306Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3188415Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3188424Z 2025-12-04T09:36:41.3188516Z Expected 0 but got 123. 2025-12-04T09:36:41.3188605Z Absolute difference: 123 2025-12-04T09:36:41.3188695Z Relative difference: inf 2025-12-04T09:36:41.3188699Z 2025-12-04T09:36:41.3188882Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3189326Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3189330Z 2025-12-04T09:36:41.3189546Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3189752Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3189854Z Traceback (most recent call last): 2025-12-04T09:36:41.3190172Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3190302Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3190719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3190890Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3191082Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3191086Z 2025-12-04T09:36:41.3191199Z Expected 0 but got 123. 2025-12-04T09:36:41.3191297Z Absolute difference: 123 2025-12-04T09:36:41.3191409Z Relative difference: inf 2025-12-04T09:36:41.3191415Z 2025-12-04T09:36:41.3191593Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3192032Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3192036Z 2025-12-04T09:36:41.3192258Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3192464Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3192571Z Traceback (most recent call last): 2025-12-04T09:36:41.3192891Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3193022Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3193440Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3193609Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3193718Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3193723Z 2025-12-04T09:36:41.3193813Z Expected 0 but got 123. 2025-12-04T09:36:41.3193903Z Absolute difference: 123 2025-12-04T09:36:41.3193992Z Relative difference: inf 2025-12-04T09:36:41.3193996Z 2025-12-04T09:36:41.3194178Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3194616Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3194698Z 2025-12-04T09:36:41.3194918Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3195129Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3195231Z Traceback (most recent call last): 2025-12-04T09:36:41.3195552Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3195683Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3196100Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3196267Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3196376Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3196381Z 2025-12-04T09:36:41.3196470Z Expected 0 but got 123. 2025-12-04T09:36:41.3196560Z Absolute difference: 123 2025-12-04T09:36:41.3196655Z Relative difference: inf 2025-12-04T09:36:41.3196658Z 2025-12-04T09:36:41.3196835Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3197282Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3197286Z 2025-12-04T09:36:41.3197503Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3197713Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3197815Z Traceback (most recent call last): 2025-12-04T09:36:41.3198133Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3198264Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3198678Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3198857Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3198966Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3198970Z 2025-12-04T09:36:41.3199059Z Expected 0 but got 123. 2025-12-04T09:36:41.3199228Z Absolute difference: 123 2025-12-04T09:36:41.3199320Z Relative difference: inf 2025-12-04T09:36:41.3199324Z 2025-12-04T09:36:41.3199504Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3199946Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3199950Z 2025-12-04T09:36:41.3200167Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3200372Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3200474Z Traceback (most recent call last): 2025-12-04T09:36:41.3200795Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3200930Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3201376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3201570Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3201679Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3201683Z 2025-12-04T09:36:41.3201770Z Expected 0 but got 123. 2025-12-04T09:36:41.3201859Z Absolute difference: 123 2025-12-04T09:36:41.3201953Z Relative difference: inf 2025-12-04T09:36:41.3201957Z 2025-12-04T09:36:41.3202134Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3202572Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3202576Z 2025-12-04T09:36:41.3202792Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3203077Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3203178Z Traceback (most recent call last): 2025-12-04T09:36:41.3203503Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3203635Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3204047Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3204217Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3204326Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3204330Z 2025-12-04T09:36:41.3204420Z Expected 0 but got 123. 2025-12-04T09:36:41.3204509Z Absolute difference: 123 2025-12-04T09:36:41.3204599Z Relative difference: inf 2025-12-04T09:36:41.3204603Z 2025-12-04T09:36:41.3204781Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3205233Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3205237Z 2025-12-04T09:36:41.3205459Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3205665Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3205766Z Traceback (most recent call last): 2025-12-04T09:36:41.3206086Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3206217Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3206630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3206799Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3206915Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3206919Z 2025-12-04T09:36:41.3207008Z Expected 0 but got 123. 2025-12-04T09:36:41.3207097Z Absolute difference: 123 2025-12-04T09:36:41.3207188Z Relative difference: inf 2025-12-04T09:36:41.3207192Z 2025-12-04T09:36:41.3207506Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3207953Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3207957Z 2025-12-04T09:36:41.3208174Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3208381Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3208483Z Traceback (most recent call last): 2025-12-04T09:36:41.3208802Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3208933Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3209351Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3209525Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3209633Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3209637Z 2025-12-04T09:36:41.3209727Z Expected 0 but got 123. 2025-12-04T09:36:41.3209815Z Absolute difference: 123 2025-12-04T09:36:41.3209904Z Relative difference: inf 2025-12-04T09:36:41.3209908Z 2025-12-04T09:36:41.3210087Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3210528Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3210532Z 2025-12-04T09:36:41.3210749Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3210955Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3211141Z Traceback (most recent call last): 2025-12-04T09:36:41.3211461Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3211600Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3212012Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3212180Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3212290Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3212294Z 2025-12-04T09:36:41.3212382Z Expected 0 but got 123. 2025-12-04T09:36:41.3212471Z Absolute difference: 123 2025-12-04T09:36:41.3212561Z Relative difference: inf 2025-12-04T09:36:41.3212564Z 2025-12-04T09:36:41.3212742Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3213179Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3213190Z 2025-12-04T09:36:41.3213406Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3213617Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3213719Z Traceback (most recent call last): 2025-12-04T09:36:41.3214038Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3214168Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3214581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3214750Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3214858Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3214862Z 2025-12-04T09:36:41.3214951Z Expected 0 but got 123. 2025-12-04T09:36:41.3215046Z Absolute difference: 123 2025-12-04T09:36:41.3215134Z Relative difference: inf 2025-12-04T09:36:41.3215138Z 2025-12-04T09:36:41.3215316Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3215865Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3215870Z 2025-12-04T09:36:41.3216086Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3216296Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3216400Z Traceback (most recent call last): 2025-12-04T09:36:41.3216718Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3216849Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3217262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3217436Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3217545Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3217549Z 2025-12-04T09:36:41.3217644Z Expected 0 but got 123. 2025-12-04T09:36:41.3217734Z Absolute difference: 123 2025-12-04T09:36:41.3217823Z Relative difference: inf 2025-12-04T09:36:41.3217827Z 2025-12-04T09:36:41.3218004Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3218445Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3218450Z 2025-12-04T09:36:41.3218665Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3218873Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3218974Z Traceback (most recent call last): 2025-12-04T09:36:41.3219375Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3219507Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3219927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3220097Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3220206Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3220210Z 2025-12-04T09:36:41.3220300Z Expected 0 but got 123. 2025-12-04T09:36:41.3220388Z Absolute difference: 123 2025-12-04T09:36:41.3220477Z Relative difference: inf 2025-12-04T09:36:41.3220481Z 2025-12-04T09:36:41.3220680Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3221154Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3221166Z 2025-12-04T09:36:41.3221378Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3221588Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3221695Z Traceback (most recent call last): 2025-12-04T09:36:41.3222018Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3222148Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3222561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3222730Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3222839Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3222843Z 2025-12-04T09:36:41.3222932Z Expected 0 but got 123. 2025-12-04T09:36:41.3223022Z Absolute difference: 123 2025-12-04T09:36:41.3223111Z Relative difference: inf 2025-12-04T09:36:41.3223119Z 2025-12-04T09:36:41.3223297Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3223815Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3223820Z 2025-12-04T09:36:41.3224037Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3224247Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3224349Z Traceback (most recent call last): 2025-12-04T09:36:41.3224669Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3224800Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3225212Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3225381Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3225495Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3225499Z 2025-12-04T09:36:41.3225585Z Expected 0 but got 123. 2025-12-04T09:36:41.3225679Z Absolute difference: 123 2025-12-04T09:36:41.3225774Z Relative difference: inf 2025-12-04T09:36:41.3225778Z 2025-12-04T09:36:41.3225955Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3226397Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3226401Z 2025-12-04T09:36:41.3226618Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3226827Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3226929Z Traceback (most recent call last): 2025-12-04T09:36:41.3227249Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3227802Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3228215Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3228389Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3228500Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3228504Z 2025-12-04T09:36:41.3228591Z Expected 0 but got 123. 2025-12-04T09:36:41.3228685Z Absolute difference: 123 2025-12-04T09:36:41.3228775Z Relative difference: inf 2025-12-04T09:36:41.3228778Z 2025-12-04T09:36:41.3228955Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3229398Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3229402Z 2025-12-04T09:36:41.3229618Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3229831Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3229933Z Traceback (most recent call last): 2025-12-04T09:36:41.3230256Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3230388Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3230801Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3230970Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3231079Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3231082Z 2025-12-04T09:36:41.3231169Z Expected 0 but got 123. 2025-12-04T09:36:41.3231260Z Absolute difference: 123 2025-12-04T09:36:41.3231349Z Relative difference: inf 2025-12-04T09:36:41.3231353Z 2025-12-04T09:36:41.3231529Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3231972Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3231976Z 2025-12-04T09:36:41.3232271Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3232482Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3232584Z Traceback (most recent call last): 2025-12-04T09:36:41.3232903Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3233034Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3233449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3233621Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3233730Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3233740Z 2025-12-04T09:36:41.3233826Z Expected 0 but got 123. 2025-12-04T09:36:41.3233919Z Absolute difference: 123 2025-12-04T09:36:41.3234008Z Relative difference: inf 2025-12-04T09:36:41.3234012Z 2025-12-04T09:36:41.3234197Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3234635Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3234639Z 2025-12-04T09:36:41.3234853Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3235064Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3235166Z Traceback (most recent call last): 2025-12-04T09:36:41.3235665Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3235797Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3236330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3236502Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3236617Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3236621Z 2025-12-04T09:36:41.3236712Z Expected 0 but got 123. 2025-12-04T09:36:41.3236804Z Absolute difference: 123 2025-12-04T09:36:41.3236893Z Relative difference: inf 2025-12-04T09:36:41.3236897Z 2025-12-04T09:36:41.3237074Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3237515Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3237519Z 2025-12-04T09:36:41.3237732Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3237940Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3238048Z Traceback (most recent call last): 2025-12-04T09:36:41.3238366Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3238503Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3238915Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3239084Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3239196Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3239200Z 2025-12-04T09:36:41.3239289Z Expected 0 but got 123. 2025-12-04T09:36:41.3239382Z Absolute difference: 123 2025-12-04T09:36:41.3239471Z Relative difference: inf 2025-12-04T09:36:41.3239475Z 2025-12-04T09:36:41.3239653Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3240091Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3240100Z 2025-12-04T09:36:41.3240314Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3240630Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3240735Z Traceback (most recent call last): 2025-12-04T09:36:41.3241051Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3241185Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3241649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3241818Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3241927Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3241931Z 2025-12-04T09:36:41.3242017Z Expected 0 but got 123. 2025-12-04T09:36:41.3242113Z Absolute difference: 123 2025-12-04T09:36:41.3242208Z Relative difference: inf 2025-12-04T09:36:41.3242212Z 2025-12-04T09:36:41.3242391Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3242844Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3242848Z 2025-12-04T09:36:41.3243061Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3243270Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3243372Z Traceback (most recent call last): 2025-12-04T09:36:41.3243689Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3243822Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3244236Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3244486Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3244595Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3244599Z 2025-12-04T09:36:41.3244686Z Expected 0 but got 123. 2025-12-04T09:36:41.3244787Z Absolute difference: 123 2025-12-04T09:36:41.3244876Z Relative difference: inf 2025-12-04T09:36:41.3244880Z 2025-12-04T09:36:41.3245056Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3245501Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3245506Z 2025-12-04T09:36:41.3245721Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3245931Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3246034Z Traceback (most recent call last): 2025-12-04T09:36:41.3246350Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3246488Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3246905Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3247079Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3247189Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3247192Z 2025-12-04T09:36:41.3247326Z Expected 0 but got 123. 2025-12-04T09:36:41.3247420Z Absolute difference: 123 2025-12-04T09:36:41.3247509Z Relative difference: inf 2025-12-04T09:36:41.3247513Z 2025-12-04T09:36:41.3247688Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3248133Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3248137Z 2025-12-04T09:36:41.3248363Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3248570Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3248672Z Traceback (most recent call last): 2025-12-04T09:36:41.3249096Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3249232Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3249645Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3249814Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3249924Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3249927Z 2025-12-04T09:36:41.3250014Z Expected 0 but got 123. 2025-12-04T09:36:41.3250108Z Absolute difference: 123 2025-12-04T09:36:41.3250197Z Relative difference: inf 2025-12-04T09:36:41.3250201Z 2025-12-04T09:36:41.3250381Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3250879Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3250888Z 2025-12-04T09:36:41.3251103Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3251314Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3251416Z Traceback (most recent call last): 2025-12-04T09:36:41.3251732Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3251866Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3252281Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3252449Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3252637Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3252641Z 2025-12-04T09:36:41.3252726Z Expected 0 but got 123. 2025-12-04T09:36:41.3252823Z Absolute difference: 123 2025-12-04T09:36:41.3252914Z Relative difference: inf 2025-12-04T09:36:41.3252923Z 2025-12-04T09:36:41.3253099Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3253544Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3253548Z 2025-12-04T09:36:41.3253762Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3253971Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3254074Z Traceback (most recent call last): 2025-12-04T09:36:41.3254393Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3254527Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3254948Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3255124Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3255234Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3255238Z 2025-12-04T09:36:41.3255328Z Expected 0 but got 123. 2025-12-04T09:36:41.3255424Z Absolute difference: 123 2025-12-04T09:36:41.3255513Z Relative difference: inf 2025-12-04T09:36:41.3255517Z 2025-12-04T09:36:41.3255692Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3256135Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3256139Z 2025-12-04T09:36:41.3256356Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3256568Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3256679Z Traceback (most recent call last): 2025-12-04T09:36:41.3256997Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3257221Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3257637Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3257805Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3257914Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3257918Z 2025-12-04T09:36:41.3258008Z Expected 0 but got 123. 2025-12-04T09:36:41.3258105Z Absolute difference: 123 2025-12-04T09:36:41.3258198Z Relative difference: inf 2025-12-04T09:36:41.3258202Z 2025-12-04T09:36:41.3258376Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3258821Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3258830Z 2025-12-04T09:36:41.3259045Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3259259Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3259361Z Traceback (most recent call last): 2025-12-04T09:36:41.3259679Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3259813Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3260226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3260396Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3260505Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3260509Z 2025-12-04T09:36:41.3260596Z Expected 0 but got 123. 2025-12-04T09:36:41.3260767Z Absolute difference: 123 2025-12-04T09:36:41.3260857Z Relative difference: inf 2025-12-04T09:36:41.3260861Z 2025-12-04T09:36:41.3261037Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3261487Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3261491Z 2025-12-04T09:36:41.3261706Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3261915Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3262019Z Traceback (most recent call last): 2025-12-04T09:36:41.3262336Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3262470Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3262883Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3263059Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3263169Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3263177Z 2025-12-04T09:36:41.3263264Z Expected 0 but got 123. 2025-12-04T09:36:41.3263359Z Absolute difference: 123 2025-12-04T09:36:41.3263454Z Relative difference: inf 2025-12-04T09:36:41.3263458Z 2025-12-04T09:36:41.3263634Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3264076Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3264080Z 2025-12-04T09:36:41.3264294Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3264502Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3264604Z Traceback (most recent call last): 2025-12-04T09:36:41.3264925Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3265058Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3265550Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3265719Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3265831Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3265836Z 2025-12-04T09:36:41.3265922Z Expected 0 but got 123. 2025-12-04T09:36:41.3266017Z Absolute difference: 123 2025-12-04T09:36:41.3266106Z Relative difference: inf 2025-12-04T09:36:41.3266110Z 2025-12-04T09:36:41.3266286Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3266728Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3266738Z 2025-12-04T09:36:41.3266953Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3267165Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3267272Z Traceback (most recent call last): 2025-12-04T09:36:41.3267590Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3267723Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3268137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3268304Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3268416Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3268420Z 2025-12-04T09:36:41.3268507Z Expected 0 but got 123. 2025-12-04T09:36:41.3268602Z Absolute difference: 123 2025-12-04T09:36:41.3268692Z Relative difference: inf 2025-12-04T09:36:41.3268775Z 2025-12-04T09:36:41.3268952Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3269397Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3269401Z 2025-12-04T09:36:41.3269615Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3269824Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3269931Z Traceback (most recent call last): 2025-12-04T09:36:41.3273564Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3273719Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3274145Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3274317Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3274444Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3274450Z 2025-12-04T09:36:41.3274538Z Expected 0 but got 123. 2025-12-04T09:36:41.3274635Z Absolute difference: 123 2025-12-04T09:36:41.3274733Z Relative difference: inf 2025-12-04T09:36:41.3274738Z 2025-12-04T09:36:41.3274915Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3275372Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3275376Z 2025-12-04T09:36:41.3275593Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3275804Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3275907Z Traceback (most recent call last): 2025-12-04T09:36:41.3276231Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3276371Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3276787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3277057Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3277178Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3277182Z 2025-12-04T09:36:41.3277270Z Expected 0 but got 123. 2025-12-04T09:36:41.3277368Z Absolute difference: 123 2025-12-04T09:36:41.3277462Z Relative difference: inf 2025-12-04T09:36:41.3277466Z 2025-12-04T09:36:41.3277643Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3278092Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3278096Z 2025-12-04T09:36:41.3278312Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3278530Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3278634Z Traceback (most recent call last): 2025-12-04T09:36:41.3278957Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3279091Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3279507Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3279675Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3279786Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3279791Z 2025-12-04T09:36:41.3279879Z Expected 0 but got 123. 2025-12-04T09:36:41.3279977Z Absolute difference: 123 2025-12-04T09:36:41.3280067Z Relative difference: inf 2025-12-04T09:36:41.3280071Z 2025-12-04T09:36:41.3280247Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3280881Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3280885Z 2025-12-04T09:36:41.3281109Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3281325Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3281440Z Traceback (most recent call last): 2025-12-04T09:36:41.3281806Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3281940Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3282357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3282525Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3282641Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3282650Z 2025-12-04T09:36:41.3282736Z Expected 0 but got 123. 2025-12-04T09:36:41.3282829Z Absolute difference: 123 2025-12-04T09:36:41.3282918Z Relative difference: inf 2025-12-04T09:36:41.3282922Z 2025-12-04T09:36:41.3283109Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3283560Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3283564Z 2025-12-04T09:36:41.3283779Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3283986Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3284095Z Traceback (most recent call last): 2025-12-04T09:36:41.3284413Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3284549Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3284966Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3285134Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3285328Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3285333Z 2025-12-04T09:36:41.3285422Z Expected 0 but got 123. 2025-12-04T09:36:41.3285512Z Absolute difference: 123 2025-12-04T09:36:41.3285605Z Relative difference: inf 2025-12-04T09:36:41.3285609Z 2025-12-04T09:36:41.3285785Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3286232Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3286236Z 2025-12-04T09:36:41.3286451Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3286660Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3286770Z Traceback (most recent call last): 2025-12-04T09:36:41.3287091Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3287235Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3287710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3287877Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3287989Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3287993Z 2025-12-04T09:36:41.3288079Z Expected 0 but got 123. 2025-12-04T09:36:41.3288169Z Absolute difference: 123 2025-12-04T09:36:41.3288259Z Relative difference: inf 2025-12-04T09:36:41.3288263Z 2025-12-04T09:36:41.3288440Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3288887Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3288974Z 2025-12-04T09:36:41.3289191Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3289406Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3289517Z Traceback (most recent call last): 2025-12-04T09:36:41.3289837Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3289973Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3290387Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3290554Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3290667Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3290671Z 2025-12-04T09:36:41.3290757Z Expected 0 but got 123. 2025-12-04T09:36:41.3290853Z Absolute difference: 123 2025-12-04T09:36:41.3290947Z Relative difference: inf 2025-12-04T09:36:41.3290950Z 2025-12-04T09:36:41.3291131Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3291582Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3291586Z 2025-12-04T09:36:41.3291801Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3292010Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3292116Z Traceback (most recent call last): 2025-12-04T09:36:41.3292433Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3292566Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3292981Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3293154Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3293269Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3293273Z 2025-12-04T09:36:41.3293440Z Expected 0 but got 123. 2025-12-04T09:36:41.3293531Z Absolute difference: 123 2025-12-04T09:36:41.3293626Z Relative difference: inf 2025-12-04T09:36:41.3293630Z 2025-12-04T09:36:41.3293806Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3294250Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3294254Z 2025-12-04T09:36:41.3294468Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3294675Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3294779Z Traceback (most recent call last): 2025-12-04T09:36:41.3295097Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3295237Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3295660Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3295828Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3295938Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3295942Z 2025-12-04T09:36:41.3296029Z Expected 0 but got 123. 2025-12-04T09:36:41.3296118Z Absolute difference: 123 2025-12-04T09:36:41.3296212Z Relative difference: inf 2025-12-04T09:36:41.3296216Z 2025-12-04T09:36:41.3296392Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3296842Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3296846Z 2025-12-04T09:36:41.3297152Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3297359Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3297463Z Traceback (most recent call last): 2025-12-04T09:36:41.3297786Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3297922Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3298337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3298504Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3298615Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3298620Z 2025-12-04T09:36:41.3298708Z Expected 0 but got 123. 2025-12-04T09:36:41.3298797Z Absolute difference: 123 2025-12-04T09:36:41.3298890Z Relative difference: inf 2025-12-04T09:36:41.3298894Z 2025-12-04T09:36:41.3299076Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3299524Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3299531Z 2025-12-04T09:36:41.3299746Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3299951Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3300056Z Traceback (most recent call last): 2025-12-04T09:36:41.3300377Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3300513Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3300933Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3301100Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3301220Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3301224Z 2025-12-04T09:36:41.3301311Z Expected 0 but got 123. 2025-12-04T09:36:41.3301403Z Absolute difference: 123 2025-12-04T09:36:41.3301496Z Relative difference: inf 2025-12-04T09:36:41.3301579Z 2025-12-04T09:36:41.3301762Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3302209Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3302213Z 2025-12-04T09:36:41.3302426Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3302634Z ________________ TestMemPool.test_mempool_empty_cache_inactive _________________ 2025-12-04T09:36:41.3302740Z Traceback (most recent call last): 2025-12-04T09:36:41.3303058Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5504, in test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3303196Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3303612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3303784Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3303896Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3303900Z 2025-12-04T09:36:41.3303986Z Expected 0 but got 123. 2025-12-04T09:36:41.3304076Z Absolute difference: 123 2025-12-04T09:36:41.3304166Z Relative difference: inf 2025-12-04T09:36:41.3304170Z 2025-12-04T09:36:41.3304346Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3304794Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3304798Z 2025-12-04T09:36:41.3305013Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3305294Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3305399Z Traceback (most recent call last): 2025-12-04T09:36:41.3305701Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3305835Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3306260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3306427Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3306539Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3306543Z 2025-12-04T09:36:41.3306630Z Expected 0 but got 123. 2025-12-04T09:36:41.3306724Z Absolute difference: 123 2025-12-04T09:36:41.3306817Z Relative difference: inf 2025-12-04T09:36:41.3306821Z 2025-12-04T09:36:41.3306996Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3307425Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3307435Z 2025-12-04T09:36:41.3307651Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3307856Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3307961Z Traceback (most recent call last): 2025-12-04T09:36:41.3308256Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3308388Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3308806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3308974Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3309089Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3309094Z 2025-12-04T09:36:41.3309181Z Expected 0 but got 123. 2025-12-04T09:36:41.3309277Z Absolute difference: 123 2025-12-04T09:36:41.3309370Z Relative difference: inf 2025-12-04T09:36:41.3309374Z 2025-12-04T09:36:41.3309550Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3310057Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3310066Z 2025-12-04T09:36:41.3310282Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3310482Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3310586Z Traceback (most recent call last): 2025-12-04T09:36:41.3310880Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3311013Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3311435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3311628Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3311752Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3311757Z 2025-12-04T09:36:41.3311866Z Expected 0 but got 123. 2025-12-04T09:36:41.3311956Z Absolute difference: 123 2025-12-04T09:36:41.3312049Z Relative difference: inf 2025-12-04T09:36:41.3312053Z 2025-12-04T09:36:41.3312231Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3312663Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3312670Z 2025-12-04T09:36:41.3312889Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3313089Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3313193Z Traceback (most recent call last): 2025-12-04T09:36:41.3313593Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3313726Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3314154Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3314322Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3314432Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3314437Z 2025-12-04T09:36:41.3314524Z Expected 0 but got 123. 2025-12-04T09:36:41.3314615Z Absolute difference: 123 2025-12-04T09:36:41.3314707Z Relative difference: inf 2025-12-04T09:36:41.3314711Z 2025-12-04T09:36:41.3314887Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3315317Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3315330Z 2025-12-04T09:36:41.3315545Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3315743Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3315851Z Traceback (most recent call last): 2025-12-04T09:36:41.3316146Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3316277Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3316694Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3316861Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3316974Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3316978Z 2025-12-04T09:36:41.3317064Z Expected 0 but got 123. 2025-12-04T09:36:41.3317155Z Absolute difference: 123 2025-12-04T09:36:41.3317246Z Relative difference: inf 2025-12-04T09:36:41.3317256Z 2025-12-04T09:36:41.3317435Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3317862Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3317952Z 2025-12-04T09:36:41.3318168Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3318368Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3318475Z Traceback (most recent call last): 2025-12-04T09:36:41.3318768Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3318899Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3319316Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3319486Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3319604Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3319608Z 2025-12-04T09:36:41.3319696Z Expected 0 but got 123. 2025-12-04T09:36:41.3319786Z Absolute difference: 123 2025-12-04T09:36:41.3319880Z Relative difference: inf 2025-12-04T09:36:41.3319889Z 2025-12-04T09:36:41.3320065Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3320491Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3320495Z 2025-12-04T09:36:41.3320713Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3320917Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3321023Z Traceback (most recent call last): 2025-12-04T09:36:41.3321315Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3321470Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3321996Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3322172Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3322282Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3322289Z 2025-12-04T09:36:41.3322375Z Expected 0 but got 123. 2025-12-04T09:36:41.3322464Z Absolute difference: 123 2025-12-04T09:36:41.3322557Z Relative difference: inf 2025-12-04T09:36:41.3322561Z 2025-12-04T09:36:41.3322735Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3323158Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3323163Z 2025-12-04T09:36:41.3323380Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3323579Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3323693Z Traceback (most recent call last): 2025-12-04T09:36:41.3323986Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3324120Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3324538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3324706Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3324816Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3324824Z 2025-12-04T09:36:41.3324911Z Expected 0 but got 123. 2025-12-04T09:36:41.3325001Z Absolute difference: 123 2025-12-04T09:36:41.3325094Z Relative difference: inf 2025-12-04T09:36:41.3325097Z 2025-12-04T09:36:41.3325272Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3325698Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3325710Z 2025-12-04T09:36:41.3325927Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3326207Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3326313Z Traceback (most recent call last): 2025-12-04T09:36:41.3326608Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3326739Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3327155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3327391Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3327502Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3327507Z 2025-12-04T09:36:41.3327596Z Expected 0 but got 123. 2025-12-04T09:36:41.3327694Z Absolute difference: 123 2025-12-04T09:36:41.3327787Z Relative difference: inf 2025-12-04T09:36:41.3327791Z 2025-12-04T09:36:41.3327966Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3328398Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3328403Z 2025-12-04T09:36:41.3328625Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3328823Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3328927Z Traceback (most recent call last): 2025-12-04T09:36:41.3329221Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3329356Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3329775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3330027Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3330136Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3330141Z 2025-12-04T09:36:41.3330238Z Expected 0 but got 123. 2025-12-04T09:36:41.3330329Z Absolute difference: 123 2025-12-04T09:36:41.3330421Z Relative difference: inf 2025-12-04T09:36:41.3330425Z 2025-12-04T09:36:41.3330601Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3331027Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3331032Z 2025-12-04T09:36:41.3331249Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3331447Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3331551Z Traceback (most recent call last): 2025-12-04T09:36:41.3331853Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3331985Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3332405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3332573Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3332683Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3332687Z 2025-12-04T09:36:41.3332777Z Expected 0 but got 123. 2025-12-04T09:36:41.3332867Z Absolute difference: 123 2025-12-04T09:36:41.3332959Z Relative difference: inf 2025-12-04T09:36:41.3332964Z 2025-12-04T09:36:41.3333140Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3333565Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3333569Z 2025-12-04T09:36:41.3333793Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3333991Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3334093Z Traceback (most recent call last): 2025-12-04T09:36:41.3334475Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3334609Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3335035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3335202Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3335490Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3335495Z 2025-12-04T09:36:41.3335585Z Expected 0 but got 123. 2025-12-04T09:36:41.3335675Z Absolute difference: 123 2025-12-04T09:36:41.3335765Z Relative difference: inf 2025-12-04T09:36:41.3335772Z 2025-12-04T09:36:41.3335953Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3336378Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3336388Z 2025-12-04T09:36:41.3336610Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3336809Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3336911Z Traceback (most recent call last): 2025-12-04T09:36:41.3337208Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3337339Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3337756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3337927Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3338161Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3338165Z 2025-12-04T09:36:41.3338256Z Expected 0 but got 123. 2025-12-04T09:36:41.3338348Z Absolute difference: 123 2025-12-04T09:36:41.3338438Z Relative difference: inf 2025-12-04T09:36:41.3338451Z 2025-12-04T09:36:41.3338628Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3339050Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3339054Z 2025-12-04T09:36:41.3339279Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3339480Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3339581Z Traceback (most recent call last): 2025-12-04T09:36:41.3339876Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3340009Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3340442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3340640Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3340765Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3340769Z 2025-12-04T09:36:41.3340859Z Expected 0 but got 123. 2025-12-04T09:36:41.3340948Z Absolute difference: 123 2025-12-04T09:36:41.3341037Z Relative difference: inf 2025-12-04T09:36:41.3341041Z 2025-12-04T09:36:41.3341219Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3341644Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3341648Z 2025-12-04T09:36:41.3341866Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3342066Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3342174Z Traceback (most recent call last): 2025-12-04T09:36:41.3342472Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3342741Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3343161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3343328Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3343438Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3343442Z 2025-12-04T09:36:41.3343533Z Expected 0 but got 123. 2025-12-04T09:36:41.3343623Z Absolute difference: 123 2025-12-04T09:36:41.3343712Z Relative difference: inf 2025-12-04T09:36:41.3343716Z 2025-12-04T09:36:41.3343892Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3344319Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3344329Z 2025-12-04T09:36:41.3344547Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3344752Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3344855Z Traceback (most recent call last): 2025-12-04T09:36:41.3345155Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3345288Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3345703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3345872Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3345982Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3345986Z 2025-12-04T09:36:41.3346075Z Expected 0 but got 123. 2025-12-04T09:36:41.3346248Z Absolute difference: 123 2025-12-04T09:36:41.3346339Z Relative difference: inf 2025-12-04T09:36:41.3346343Z 2025-12-04T09:36:41.3346521Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3346959Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3346963Z 2025-12-04T09:36:41.3347183Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3347382Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3347482Z Traceback (most recent call last): 2025-12-04T09:36:41.3347778Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3347915Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3348329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3348506Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3348620Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3348624Z 2025-12-04T09:36:41.3348717Z Expected 0 but got 123. 2025-12-04T09:36:41.3348807Z Absolute difference: 123 2025-12-04T09:36:41.3348896Z Relative difference: inf 2025-12-04T09:36:41.3348900Z 2025-12-04T09:36:41.3349078Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3349506Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3349510Z 2025-12-04T09:36:41.3349727Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3349927Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3350028Z Traceback (most recent call last): 2025-12-04T09:36:41.3350325Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3350461Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3350960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3351130Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3351239Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3351243Z 2025-12-04T09:36:41.3351332Z Expected 0 but got 123. 2025-12-04T09:36:41.3351422Z Absolute difference: 123 2025-12-04T09:36:41.3351511Z Relative difference: inf 2025-12-04T09:36:41.3351515Z 2025-12-04T09:36:41.3351696Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3352123Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3352127Z 2025-12-04T09:36:41.3352344Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3352548Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3352650Z Traceback (most recent call last): 2025-12-04T09:36:41.3352952Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3353085Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3353500Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3353672Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3353783Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3353787Z 2025-12-04T09:36:41.3353877Z Expected 0 but got 123. 2025-12-04T09:36:41.3353967Z Absolute difference: 123 2025-12-04T09:36:41.3354057Z Relative difference: inf 2025-12-04T09:36:41.3354061Z 2025-12-04T09:36:41.3354238Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3354737Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3354741Z 2025-12-04T09:36:41.3354968Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3355167Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3355268Z Traceback (most recent call last): 2025-12-04T09:36:41.3355565Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3355697Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3356111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3356287Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3356397Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3356408Z 2025-12-04T09:36:41.3356502Z Expected 0 but got 123. 2025-12-04T09:36:41.3356592Z Absolute difference: 123 2025-12-04T09:36:41.3356682Z Relative difference: inf 2025-12-04T09:36:41.3356686Z 2025-12-04T09:36:41.3356870Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3357293Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3357297Z 2025-12-04T09:36:41.3357516Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3357715Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3357818Z Traceback (most recent call last): 2025-12-04T09:36:41.3358113Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3358246Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3358670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3358839Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3359030Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3359035Z 2025-12-04T09:36:41.3359124Z Expected 0 but got 123. 2025-12-04T09:36:41.3359214Z Absolute difference: 123 2025-12-04T09:36:41.3359304Z Relative difference: inf 2025-12-04T09:36:41.3359308Z 2025-12-04T09:36:41.3359485Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3359911Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3359915Z 2025-12-04T09:36:41.3360134Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3360332Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3360443Z Traceback (most recent call last): 2025-12-04T09:36:41.3360743Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3360879Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3361301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3361499Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3361623Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3361627Z 2025-12-04T09:36:41.3361716Z Expected 0 but got 123. 2025-12-04T09:36:41.3361806Z Absolute difference: 123 2025-12-04T09:36:41.3361895Z Relative difference: inf 2025-12-04T09:36:41.3361899Z 2025-12-04T09:36:41.3362078Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3362503Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3362586Z 2025-12-04T09:36:41.3362806Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3363011Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3363113Z Traceback (most recent call last): 2025-12-04T09:36:41.3363408Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3363539Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3363955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3364126Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3364236Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3364240Z 2025-12-04T09:36:41.3364328Z Expected 0 but got 123. 2025-12-04T09:36:41.3364418Z Absolute difference: 123 2025-12-04T09:36:41.3364513Z Relative difference: inf 2025-12-04T09:36:41.3364517Z 2025-12-04T09:36:41.3364693Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3365123Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3365127Z 2025-12-04T09:36:41.3365345Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3365543Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3365644Z Traceback (most recent call last): 2025-12-04T09:36:41.3365938Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3366068Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3366481Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3366657Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3366770Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3366774Z 2025-12-04T09:36:41.3366863Z Expected 0 but got 123. 2025-12-04T09:36:41.3367030Z Absolute difference: 123 2025-12-04T09:36:41.3367120Z Relative difference: inf 2025-12-04T09:36:41.3367124Z 2025-12-04T09:36:41.3367348Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3367773Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3367777Z 2025-12-04T09:36:41.3367991Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3368192Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3368294Z Traceback (most recent call last): 2025-12-04T09:36:41.3368592Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3368729Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3369146Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3369316Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3369424Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3369428Z 2025-12-04T09:36:41.3369520Z Expected 0 but got 123. 2025-12-04T09:36:41.3369610Z Absolute difference: 123 2025-12-04T09:36:41.3369700Z Relative difference: inf 2025-12-04T09:36:41.3369703Z 2025-12-04T09:36:41.3369881Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3370306Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3370310Z 2025-12-04T09:36:41.3370524Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3370809Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3370912Z Traceback (most recent call last): 2025-12-04T09:36:41.3371239Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3371394Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3371808Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3371977Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3372086Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3372090Z 2025-12-04T09:36:41.3372176Z Expected 0 but got 123. 2025-12-04T09:36:41.3372267Z Absolute difference: 123 2025-12-04T09:36:41.3372355Z Relative difference: inf 2025-12-04T09:36:41.3372359Z 2025-12-04T09:36:41.3372536Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3372966Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3372970Z 2025-12-04T09:36:41.3373188Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3373390Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3373495Z Traceback (most recent call last): 2025-12-04T09:36:41.3373792Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3373923Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3374335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3374504Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3374613Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3374622Z 2025-12-04T09:36:41.3374708Z Expected 0 but got 123. 2025-12-04T09:36:41.3374801Z Absolute difference: 123 2025-12-04T09:36:41.3374890Z Relative difference: inf 2025-12-04T09:36:41.3374894Z 2025-12-04T09:36:41.3375177Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3375605Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3375609Z 2025-12-04T09:36:41.3375823Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3376024Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3376126Z Traceback (most recent call last): 2025-12-04T09:36:41.3376421Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3376551Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3376969Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3377138Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3377251Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3377256Z 2025-12-04T09:36:41.3377342Z Expected 0 but got 123. 2025-12-04T09:36:41.3377433Z Absolute difference: 123 2025-12-04T09:36:41.3377523Z Relative difference: inf 2025-12-04T09:36:41.3377527Z 2025-12-04T09:36:41.3377704Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3378129Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3378133Z 2025-12-04T09:36:41.3378350Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3378551Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3378731Z Traceback (most recent call last): 2025-12-04T09:36:41.3379026Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3379162Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3379575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3379747Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3379856Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3379860Z 2025-12-04T09:36:41.3379946Z Expected 0 but got 123. 2025-12-04T09:36:41.3380041Z Absolute difference: 123 2025-12-04T09:36:41.3380129Z Relative difference: inf 2025-12-04T09:36:41.3380133Z 2025-12-04T09:36:41.3380311Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3380786Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3380796Z 2025-12-04T09:36:41.3381011Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3381215Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3381316Z Traceback (most recent call last): 2025-12-04T09:36:41.3381609Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3381743Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3382156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3382328Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3382436Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3382441Z 2025-12-04T09:36:41.3382526Z Expected 0 but got 123. 2025-12-04T09:36:41.3382618Z Absolute difference: 123 2025-12-04T09:36:41.3382711Z Relative difference: inf 2025-12-04T09:36:41.3382715Z 2025-12-04T09:36:41.3382891Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3383395Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3383400Z 2025-12-04T09:36:41.3383614Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3383815Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3383917Z Traceback (most recent call last): 2025-12-04T09:36:41.3384209Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3384343Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3384756Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3384930Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3385038Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3385042Z 2025-12-04T09:36:41.3385128Z Expected 0 but got 123. 2025-12-04T09:36:41.3385226Z Absolute difference: 123 2025-12-04T09:36:41.3385317Z Relative difference: inf 2025-12-04T09:36:41.3385320Z 2025-12-04T09:36:41.3385497Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3385918Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3385922Z 2025-12-04T09:36:41.3386137Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3386343Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3386444Z Traceback (most recent call last): 2025-12-04T09:36:41.3386735Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3386948Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3387367Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3387536Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3387645Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3387649Z 2025-12-04T09:36:41.3387737Z Expected 0 but got 123. 2025-12-04T09:36:41.3387828Z Absolute difference: 123 2025-12-04T09:36:41.3387917Z Relative difference: inf 2025-12-04T09:36:41.3387921Z 2025-12-04T09:36:41.3388099Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3388525Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3388529Z 2025-12-04T09:36:41.3388743Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3388951Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3389052Z Traceback (most recent call last): 2025-12-04T09:36:41.3389349Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3389485Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3389900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3390068Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3390177Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3390181Z 2025-12-04T09:36:41.3390267Z Expected 0 but got 123. 2025-12-04T09:36:41.3390358Z Absolute difference: 123 2025-12-04T09:36:41.3390447Z Relative difference: inf 2025-12-04T09:36:41.3390451Z 2025-12-04T09:36:41.3390624Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3391062Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3391066Z 2025-12-04T09:36:41.3391361Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3391564Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3391665Z Traceback (most recent call last): 2025-12-04T09:36:41.3391957Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3392091Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3392504Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3392675Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3392783Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3392792Z 2025-12-04T09:36:41.3392879Z Expected 0 but got 123. 2025-12-04T09:36:41.3392970Z Absolute difference: 123 2025-12-04T09:36:41.3393060Z Relative difference: inf 2025-12-04T09:36:41.3393064Z 2025-12-04T09:36:41.3393243Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3393671Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3393675Z 2025-12-04T09:36:41.3393888Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3394089Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3394191Z Traceback (most recent call last): 2025-12-04T09:36:41.3394484Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3394617Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3395111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3395283Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3395398Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3395402Z 2025-12-04T09:36:41.3395489Z Expected 0 but got 123. 2025-12-04T09:36:41.3395586Z Absolute difference: 123 2025-12-04T09:36:41.3395675Z Relative difference: inf 2025-12-04T09:36:41.3395679Z 2025-12-04T09:36:41.3395853Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3396281Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3396285Z 2025-12-04T09:36:41.3396498Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3396699Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3396807Z Traceback (most recent call last): 2025-12-04T09:36:41.3397100Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3397233Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3397650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3397820Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3397929Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3397933Z 2025-12-04T09:36:41.3398019Z Expected 0 but got 123. 2025-12-04T09:36:41.3398111Z Absolute difference: 123 2025-12-04T09:36:41.3398200Z Relative difference: inf 2025-12-04T09:36:41.3398204Z 2025-12-04T09:36:41.3398379Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3398808Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3398818Z 2025-12-04T09:36:41.3399033Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3399312Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3399416Z Traceback (most recent call last): 2025-12-04T09:36:41.3399714Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3399848Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3400261Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3400427Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3400538Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3400542Z 2025-12-04T09:36:41.3400628Z Expected 0 but got 123. 2025-12-04T09:36:41.3400722Z Absolute difference: 123 2025-12-04T09:36:41.3400815Z Relative difference: inf 2025-12-04T09:36:41.3400819Z 2025-12-04T09:36:41.3400993Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3401425Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3401430Z 2025-12-04T09:36:41.3401643Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3401844Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3401948Z Traceback (most recent call last): 2025-12-04T09:36:41.3402239Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3402373Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3402787Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3403060Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3403173Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3403177Z 2025-12-04T09:36:41.3403263Z Expected 0 but got 123. 2025-12-04T09:36:41.3403361Z Absolute difference: 123 2025-12-04T09:36:41.3403454Z Relative difference: inf 2025-12-04T09:36:41.3403458Z 2025-12-04T09:36:41.3403632Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3404060Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3404064Z 2025-12-04T09:36:41.3404278Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3404480Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3404581Z Traceback (most recent call last): 2025-12-04T09:36:41.3404874Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3405018Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3405432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3405603Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3405715Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3405719Z 2025-12-04T09:36:41.3405804Z Expected 0 but got 123. 2025-12-04T09:36:41.3405896Z Absolute difference: 123 2025-12-04T09:36:41.3405985Z Relative difference: inf 2025-12-04T09:36:41.3405989Z 2025-12-04T09:36:41.3406163Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3406592Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3406596Z 2025-12-04T09:36:41.3406812Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3407022Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3407123Z Traceback (most recent call last): 2025-12-04T09:36:41.3407561Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3407701Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3408116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3408282Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3408393Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3408397Z 2025-12-04T09:36:41.3408483Z Expected 0 but got 123. 2025-12-04T09:36:41.3408576Z Absolute difference: 123 2025-12-04T09:36:41.3408666Z Relative difference: inf 2025-12-04T09:36:41.3408669Z 2025-12-04T09:36:41.3408842Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3409280Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3409284Z 2025-12-04T09:36:41.3409506Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3409707Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3409808Z Traceback (most recent call last): 2025-12-04T09:36:41.3410099Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3410232Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3410644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3410811Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3410923Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3411087Z 2025-12-04T09:36:41.3411174Z Expected 0 but got 123. 2025-12-04T09:36:41.3411269Z Absolute difference: 123 2025-12-04T09:36:41.3411358Z Relative difference: inf 2025-12-04T09:36:41.3411362Z 2025-12-04T09:36:41.3411542Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3411971Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3411975Z 2025-12-04T09:36:41.3412190Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3412389Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3412493Z Traceback (most recent call last): 2025-12-04T09:36:41.3412788Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3412922Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3413334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3413507Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3413625Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3413629Z 2025-12-04T09:36:41.3413714Z Expected 0 but got 123. 2025-12-04T09:36:41.3413807Z Absolute difference: 123 2025-12-04T09:36:41.3413896Z Relative difference: inf 2025-12-04T09:36:41.3413900Z 2025-12-04T09:36:41.3414074Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3414505Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3414509Z 2025-12-04T09:36:41.3414722Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3414921Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3415030Z Traceback (most recent call last): 2025-12-04T09:36:41.3415322Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3415456Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3415951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3416121Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3416235Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3416239Z 2025-12-04T09:36:41.3416325Z Expected 0 but got 123. 2025-12-04T09:36:41.3416415Z Absolute difference: 123 2025-12-04T09:36:41.3416506Z Relative difference: inf 2025-12-04T09:36:41.3416509Z 2025-12-04T09:36:41.3416683Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3417107Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3417116Z 2025-12-04T09:36:41.3417330Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3417529Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3417637Z Traceback (most recent call last): 2025-12-04T09:36:41.3417929Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3418062Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3418475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3418642Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3418754Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3418758Z 2025-12-04T09:36:41.3418844Z Expected 0 but got 123. 2025-12-04T09:36:41.3418934Z Absolute difference: 123 2025-12-04T09:36:41.3419027Z Relative difference: inf 2025-12-04T09:36:41.3419109Z 2025-12-04T09:36:41.3419287Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3419722Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3419726Z 2025-12-04T09:36:41.3419939Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3420137Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3420241Z Traceback (most recent call last): 2025-12-04T09:36:41.3420534Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3420667Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3421116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3421300Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3421418Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3421422Z 2025-12-04T09:36:41.3421508Z Expected 0 but got 123. 2025-12-04T09:36:41.3421598Z Absolute difference: 123 2025-12-04T09:36:41.3421693Z Relative difference: inf 2025-12-04T09:36:41.3421697Z 2025-12-04T09:36:41.3421872Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3425550Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3425558Z 2025-12-04T09:36:41.3425795Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3425997Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3426104Z Traceback (most recent call last): 2025-12-04T09:36:41.3426401Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3426550Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3426970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3427248Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3427363Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3427367Z 2025-12-04T09:36:41.3427455Z Expected 0 but got 123. 2025-12-04T09:36:41.3427546Z Absolute difference: 123 2025-12-04T09:36:41.3427639Z Relative difference: inf 2025-12-04T09:36:41.3427643Z 2025-12-04T09:36:41.3427819Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3428248Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3428252Z 2025-12-04T09:36:41.3428467Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3428669Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3428772Z Traceback (most recent call last): 2025-12-04T09:36:41.3429073Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3429207Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3429624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3429797Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3429908Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3429912Z 2025-12-04T09:36:41.3429998Z Expected 0 but got 123. 2025-12-04T09:36:41.3430087Z Absolute difference: 123 2025-12-04T09:36:41.3430181Z Relative difference: inf 2025-12-04T09:36:41.3430185Z 2025-12-04T09:36:41.3430358Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3430916Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3430921Z 2025-12-04T09:36:41.3431140Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3431339Z ___________________ TestMemPool.test_mempool_with_allocator ____________________ 2025-12-04T09:36:41.3431443Z Traceback (most recent call last): 2025-12-04T09:36:41.3431736Z File "/var/lib/jenkins/workspace/test/test_cuda.py", line 5546, in test_mempool_with_allocator 2025-12-04T09:36:41.3431870Z self.assertEqual(called_dummy_alloc.value, 0) 2025-12-04T09:36:41.3432288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:36:41.3432454Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:36:41.3432565Z AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3432569Z 2025-12-04T09:36:41.3432664Z Expected 0 but got 123. 2025-12-04T09:36:41.3432753Z Absolute difference: 123 2025-12-04T09:36:41.3432851Z Relative difference: inf 2025-12-04T09:36:41.3432854Z 2025-12-04T09:36:41.3433038Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3433467Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3433471Z 2025-12-04T09:36:41.3433685Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3434118Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda/test_cuda-be50f111eb28e9fa.xml - 2025-12-04T09:36:41.3434263Z =========================== short test summary info ============================ 2025-12-04T09:36:41.3434783Z FAILED [0.0601s] test_cuda.py::TestCudaMallocAsync::test_memory_snapshot - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpkd1nxy_c.pl' 2025-12-04T09:36:41.3435161Z FAILED [9.4466s] test_cuda.py::TestCudaMallocAsync::test_memory_snapshot_with_cpp - AssertionError: False is not true 2025-12-04T09:36:41.3435166Z 2025-12-04T09:36:41.3435738Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3436207Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCudaMallocAsync.test_memory_snapshot_with_cpp 2025-12-04T09:36:41.3436212Z 2025-12-04T09:36:41.3436431Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3436780Z FAILED [0.0024s] test_cuda.py::TestMemPool::test_mempool_ctx_multithread - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3436785Z 2025-12-04T09:36:41.3436873Z Expected 0 but got 1. 2025-12-04T09:36:41.3436961Z Absolute difference: 1 2025-12-04T09:36:41.3437051Z Relative difference: inf 2025-12-04T09:36:41.3437160Z Expected empty pool in the beginning 2025-12-04T09:36:41.3437164Z 2025-12-04T09:36:41.3437345Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3437774Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_ctx_multithread 2025-12-04T09:36:41.3437790Z 2025-12-04T09:36:41.3438003Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3438304Z FAILED [0.1553s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3438309Z 2025-12-04T09:36:41.3438400Z Expected 196608 but got 0. 2025-12-04T09:36:41.3438492Z Absolute difference: 196608 2025-12-04T09:36:41.3438581Z Relative difference: 1.0 2025-12-04T09:36:41.3438586Z 2025-12-04T09:36:41.3438767Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3439158Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3439294Z 2025-12-04T09:36:41.3439511Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3439811Z FAILED [0.1540s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3439816Z 2025-12-04T09:36:41.3439909Z Expected 196608 but got 0. 2025-12-04T09:36:41.3440003Z Absolute difference: 196608 2025-12-04T09:36:41.3440092Z Relative difference: 1.0 2025-12-04T09:36:41.3440096Z 2025-12-04T09:36:41.3440272Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3440688Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3440693Z 2025-12-04T09:36:41.3440930Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3441234Z FAILED [0.1538s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3441238Z 2025-12-04T09:36:41.3441335Z Expected 196608 but got 0. 2025-12-04T09:36:41.3441428Z Absolute difference: 196608 2025-12-04T09:36:41.3441517Z Relative difference: 1.0 2025-12-04T09:36:41.3441521Z 2025-12-04T09:36:41.3441700Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3442092Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3442096Z 2025-12-04T09:36:41.3442311Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3442609Z FAILED [0.1538s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3442620Z 2025-12-04T09:36:41.3442707Z Expected 196608 but got 0. 2025-12-04T09:36:41.3442798Z Absolute difference: 196608 2025-12-04T09:36:41.3442889Z Relative difference: 1.0 2025-12-04T09:36:41.3442893Z 2025-12-04T09:36:41.3443067Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3443463Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3443468Z 2025-12-04T09:36:41.3443762Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3444067Z FAILED [0.1538s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3444072Z 2025-12-04T09:36:41.3444162Z Expected 196608 but got 0. 2025-12-04T09:36:41.3444254Z Absolute difference: 196608 2025-12-04T09:36:41.3444343Z Relative difference: 1.0 2025-12-04T09:36:41.3444347Z 2025-12-04T09:36:41.3444523Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3444912Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3444917Z 2025-12-04T09:36:41.3445131Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3445437Z FAILED [0.1541s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3445441Z 2025-12-04T09:36:41.3445528Z Expected 196608 but got 0. 2025-12-04T09:36:41.3445626Z Absolute difference: 196608 2025-12-04T09:36:41.3445716Z Relative difference: 1.0 2025-12-04T09:36:41.3445720Z 2025-12-04T09:36:41.3445900Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3446288Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3446292Z 2025-12-04T09:36:41.3446503Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3446810Z FAILED [0.1540s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3446814Z 2025-12-04T09:36:41.3446901Z Expected 196608 but got 0. 2025-12-04T09:36:41.3446992Z Absolute difference: 196608 2025-12-04T09:36:41.3447163Z Relative difference: 1.0 2025-12-04T09:36:41.3447167Z 2025-12-04T09:36:41.3447391Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3447795Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3447799Z 2025-12-04T09:36:41.3448014Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3448316Z FAILED [0.1550s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3448324Z 2025-12-04T09:36:41.3448412Z Expected 196608 but got 0. 2025-12-04T09:36:41.3448504Z Absolute difference: 196608 2025-12-04T09:36:41.3448599Z Relative difference: 1.0 2025-12-04T09:36:41.3448603Z 2025-12-04T09:36:41.3448777Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3449167Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3449177Z 2025-12-04T09:36:41.3449393Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3449697Z FAILED [0.1549s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3449702Z 2025-12-04T09:36:41.3449797Z Expected 196608 but got 0. 2025-12-04T09:36:41.3449888Z Absolute difference: 196608 2025-12-04T09:36:41.3449978Z Relative difference: 1.0 2025-12-04T09:36:41.3449982Z 2025-12-04T09:36:41.3450159Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3450548Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3450552Z 2025-12-04T09:36:41.3450768Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3451069Z FAILED [0.1547s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3451079Z 2025-12-04T09:36:41.3451166Z Expected 196608 but got 0. 2025-12-04T09:36:41.3451261Z Absolute difference: 196608 2025-12-04T09:36:41.3451435Z Relative difference: 1.0 2025-12-04T09:36:41.3451439Z 2025-12-04T09:36:41.3451614Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3452007Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3452011Z 2025-12-04T09:36:41.3452225Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3452530Z FAILED [0.1552s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3452534Z 2025-12-04T09:36:41.3452622Z Expected 196608 but got 0. 2025-12-04T09:36:41.3452715Z Absolute difference: 196608 2025-12-04T09:36:41.3452808Z Relative difference: 1.0 2025-12-04T09:36:41.3452817Z 2025-12-04T09:36:41.3452992Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3453392Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3453396Z 2025-12-04T09:36:41.3453614Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3453913Z FAILED [0.1553s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3453918Z 2025-12-04T09:36:41.3454009Z Expected 196608 but got 0. 2025-12-04T09:36:41.3454100Z Absolute difference: 196608 2025-12-04T09:36:41.3454195Z Relative difference: 1.0 2025-12-04T09:36:41.3454199Z 2025-12-04T09:36:41.3454373Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3454765Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3454846Z 2025-12-04T09:36:41.3455062Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3455369Z FAILED [0.1545s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3455374Z 2025-12-04T09:36:41.3455465Z Expected 196608 but got 0. 2025-12-04T09:36:41.3455556Z Absolute difference: 196608 2025-12-04T09:36:41.3455646Z Relative difference: 1.0 2025-12-04T09:36:41.3455650Z 2025-12-04T09:36:41.3455827Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3456218Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3456223Z 2025-12-04T09:36:41.3456435Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3456742Z FAILED [0.1550s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3456754Z 2025-12-04T09:36:41.3456842Z Expected 196608 but got 0. 2025-12-04T09:36:41.3456936Z Absolute difference: 196608 2025-12-04T09:36:41.3457025Z Relative difference: 1.0 2025-12-04T09:36:41.3457030Z 2025-12-04T09:36:41.3457207Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3457604Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3457608Z 2025-12-04T09:36:41.3457819Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3458120Z FAILED [0.1563s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3458124Z 2025-12-04T09:36:41.3458211Z Expected 196608 but got 0. 2025-12-04T09:36:41.3458302Z Absolute difference: 196608 2025-12-04T09:36:41.3458396Z Relative difference: 1.0 2025-12-04T09:36:41.3458400Z 2025-12-04T09:36:41.3458573Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3458975Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3458980Z 2025-12-04T09:36:41.3459272Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3459576Z FAILED [0.1548s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3459581Z 2025-12-04T09:36:41.3459672Z Expected 196608 but got 0. 2025-12-04T09:36:41.3459763Z Absolute difference: 196608 2025-12-04T09:36:41.3459853Z Relative difference: 1.0 2025-12-04T09:36:41.3459859Z 2025-12-04T09:36:41.3460033Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3460423Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3460427Z 2025-12-04T09:36:41.3460642Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3460948Z FAILED [0.1548s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3460952Z 2025-12-04T09:36:41.3461047Z Expected 196608 but got 0. 2025-12-04T09:36:41.3461139Z Absolute difference: 196608 2025-12-04T09:36:41.3461229Z Relative difference: 1.0 2025-12-04T09:36:41.3461233Z 2025-12-04T09:36:41.3461411Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3461801Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3461805Z 2025-12-04T09:36:41.3462020Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3462325Z FAILED [0.1543s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3462329Z 2025-12-04T09:36:41.3462417Z Expected 196608 but got 0. 2025-12-04T09:36:41.3462590Z Absolute difference: 196608 2025-12-04T09:36:41.3462680Z Relative difference: 1.0 2025-12-04T09:36:41.3462684Z 2025-12-04T09:36:41.3462859Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3463266Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3463270Z 2025-12-04T09:36:41.3463482Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3463789Z FAILED [0.1546s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3463793Z 2025-12-04T09:36:41.3463885Z Expected 196608 but got 0. 2025-12-04T09:36:41.3463977Z Absolute difference: 196608 2025-12-04T09:36:41.3464070Z Relative difference: 1.0 2025-12-04T09:36:41.3464074Z 2025-12-04T09:36:41.3464248Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3464639Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3464653Z 2025-12-04T09:36:41.3464865Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3465171Z FAILED [0.1554s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3465175Z 2025-12-04T09:36:41.3465273Z Expected 196608 but got 0. 2025-12-04T09:36:41.3465365Z Absolute difference: 196608 2025-12-04T09:36:41.3465455Z Relative difference: 1.0 2025-12-04T09:36:41.3465459Z 2025-12-04T09:36:41.3465637Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3466028Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3466032Z 2025-12-04T09:36:41.3466247Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3466551Z FAILED [0.1550s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3466555Z 2025-12-04T09:36:41.3466644Z Expected 196608 but got 0. 2025-12-04T09:36:41.3466738Z Absolute difference: 196608 2025-12-04T09:36:41.3466931Z Relative difference: 1.0 2025-12-04T09:36:41.3466935Z 2025-12-04T09:36:41.3467115Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3467504Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3467508Z 2025-12-04T09:36:41.3467723Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3468032Z FAILED [0.1555s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3468036Z 2025-12-04T09:36:41.3468124Z Expected 196608 but got 0. 2025-12-04T09:36:41.3468219Z Absolute difference: 196608 2025-12-04T09:36:41.3468316Z Relative difference: 1.0 2025-12-04T09:36:41.3468320Z 2025-12-04T09:36:41.3468494Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3468892Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3468896Z 2025-12-04T09:36:41.3469108Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3469413Z FAILED [0.1546s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3469416Z 2025-12-04T09:36:41.3469505Z Expected 196608 but got 0. 2025-12-04T09:36:41.3469597Z Absolute difference: 196608 2025-12-04T09:36:41.3469692Z Relative difference: 1.0 2025-12-04T09:36:41.3469696Z 2025-12-04T09:36:41.3469870Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3470262Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3470355Z 2025-12-04T09:36:41.3470585Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3470929Z FAILED [0.1543s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3470933Z 2025-12-04T09:36:41.3471023Z Expected 196608 but got 0. 2025-12-04T09:36:41.3471120Z Absolute difference: 196608 2025-12-04T09:36:41.3471210Z Relative difference: 1.0 2025-12-04T09:36:41.3471214Z 2025-12-04T09:36:41.3471392Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3471784Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3471788Z 2025-12-04T09:36:41.3472003Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3472301Z FAILED [0.1546s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3472310Z 2025-12-04T09:36:41.3472398Z Expected 196608 but got 0. 2025-12-04T09:36:41.3472492Z Absolute difference: 196608 2025-12-04T09:36:41.3472584Z Relative difference: 1.0 2025-12-04T09:36:41.3472588Z 2025-12-04T09:36:41.3472770Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3473162Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3473166Z 2025-12-04T09:36:41.3473377Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3473682Z FAILED [0.1546s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3473686Z 2025-12-04T09:36:41.3473774Z Expected 196608 but got 0. 2025-12-04T09:36:41.3473871Z Absolute difference: 196608 2025-12-04T09:36:41.3473961Z Relative difference: 1.0 2025-12-04T09:36:41.3473964Z 2025-12-04T09:36:41.3474143Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3474535Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3474539Z 2025-12-04T09:36:41.3474829Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3475134Z FAILED [0.1549s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3475141Z 2025-12-04T09:36:41.3475228Z Expected 196608 but got 0. 2025-12-04T09:36:41.3475321Z Absolute difference: 196608 2025-12-04T09:36:41.3475413Z Relative difference: 1.0 2025-12-04T09:36:41.3475417Z 2025-12-04T09:36:41.3475592Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3475980Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3475985Z 2025-12-04T09:36:41.3476207Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3476507Z FAILED [0.1547s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3476511Z 2025-12-04T09:36:41.3476608Z Expected 196608 but got 0. 2025-12-04T09:36:41.3476698Z Absolute difference: 196608 2025-12-04T09:36:41.3476789Z Relative difference: 1.0 2025-12-04T09:36:41.3476793Z 2025-12-04T09:36:41.3476975Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3477367Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3477371Z 2025-12-04T09:36:41.3477586Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3477888Z FAILED [0.1554s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3477892Z 2025-12-04T09:36:41.3478060Z Expected 196608 but got 0. 2025-12-04T09:36:41.3478153Z Absolute difference: 196608 2025-12-04T09:36:41.3478242Z Relative difference: 1.0 2025-12-04T09:36:41.3478246Z 2025-12-04T09:36:41.3478426Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3478820Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3478824Z 2025-12-04T09:36:41.3479036Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3479340Z FAILED [0.1547s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3479344Z 2025-12-04T09:36:41.3479431Z Expected 196608 but got 0. 2025-12-04T09:36:41.3479523Z Absolute difference: 196608 2025-12-04T09:36:41.3479616Z Relative difference: 1.0 2025-12-04T09:36:41.3479620Z 2025-12-04T09:36:41.3479795Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3480192Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3480196Z 2025-12-04T09:36:41.3480412Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3480711Z FAILED [0.1543s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3480718Z 2025-12-04T09:36:41.3480805Z Expected 196608 but got 0. 2025-12-04T09:36:41.3480896Z Absolute difference: 196608 2025-12-04T09:36:41.3480988Z Relative difference: 1.0 2025-12-04T09:36:41.3480992Z 2025-12-04T09:36:41.3481165Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3481553Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3481558Z 2025-12-04T09:36:41.3481773Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3482084Z FAILED [0.1547s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3482088Z 2025-12-04T09:36:41.3482179Z Expected 196608 but got 0. 2025-12-04T09:36:41.3482349Z Absolute difference: 196608 2025-12-04T09:36:41.3482440Z Relative difference: 1.0 2025-12-04T09:36:41.3482444Z 2025-12-04T09:36:41.3482621Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3483014Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3483019Z 2025-12-04T09:36:41.3483233Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3483536Z FAILED [0.1547s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3483540Z 2025-12-04T09:36:41.3483627Z Expected 196608 but got 0. 2025-12-04T09:36:41.3483723Z Absolute difference: 196608 2025-12-04T09:36:41.3483819Z Relative difference: 1.0 2025-12-04T09:36:41.3483822Z 2025-12-04T09:36:41.3483996Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3484392Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3484397Z 2025-12-04T09:36:41.3484610Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3484911Z FAILED [0.1549s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3484915Z 2025-12-04T09:36:41.3485002Z Expected 196608 but got 0. 2025-12-04T09:36:41.3485093Z Absolute difference: 196608 2025-12-04T09:36:41.3485186Z Relative difference: 1.0 2025-12-04T09:36:41.3485190Z 2025-12-04T09:36:41.3485364Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3485755Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3485840Z 2025-12-04T09:36:41.3486055Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3486363Z FAILED [0.1545s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3486367Z 2025-12-04T09:36:41.3486458Z Expected 196608 but got 0. 2025-12-04T09:36:41.3486549Z Absolute difference: 196608 2025-12-04T09:36:41.3486641Z Relative difference: 1.0 2025-12-04T09:36:41.3486645Z 2025-12-04T09:36:41.3486824Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3487214Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3487218Z 2025-12-04T09:36:41.3487536Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3487839Z FAILED [0.1554s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3487850Z 2025-12-04T09:36:41.3487940Z Expected 196608 but got 0. 2025-12-04T09:36:41.3488031Z Absolute difference: 196608 2025-12-04T09:36:41.3488121Z Relative difference: 1.0 2025-12-04T09:36:41.3488130Z 2025-12-04T09:36:41.3488309Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3488698Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3488702Z 2025-12-04T09:36:41.3488919Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3489225Z FAILED [0.1544s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3489229Z 2025-12-04T09:36:41.3489316Z Expected 196608 but got 0. 2025-12-04T09:36:41.3489410Z Absolute difference: 196608 2025-12-04T09:36:41.3489500Z Relative difference: 1.0 2025-12-04T09:36:41.3489509Z 2025-12-04T09:36:41.3489684Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3490157Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3490161Z 2025-12-04T09:36:41.3490376Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3490684Z FAILED [0.1546s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3490688Z 2025-12-04T09:36:41.3490775Z Expected 196608 but got 0. 2025-12-04T09:36:41.3490870Z Absolute difference: 196608 2025-12-04T09:36:41.3490962Z Relative difference: 1.0 2025-12-04T09:36:41.3490965Z 2025-12-04T09:36:41.3491137Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3491528Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3491537Z 2025-12-04T09:36:41.3491749Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3492047Z FAILED [0.1554s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3492059Z 2025-12-04T09:36:41.3492149Z Expected 196608 but got 0. 2025-12-04T09:36:41.3492240Z Absolute difference: 196608 2025-12-04T09:36:41.3492332Z Relative difference: 1.0 2025-12-04T09:36:41.3492339Z 2025-12-04T09:36:41.3492512Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3492900Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3492905Z 2025-12-04T09:36:41.3493119Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3493423Z FAILED [0.1547s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3493503Z 2025-12-04T09:36:41.3493594Z Expected 196608 but got 0. 2025-12-04T09:36:41.3493684Z Absolute difference: 196608 2025-12-04T09:36:41.3493772Z Relative difference: 1.0 2025-12-04T09:36:41.3493776Z 2025-12-04T09:36:41.3493956Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3494346Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3494350Z 2025-12-04T09:36:41.3494561Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3494866Z FAILED [0.1546s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3494870Z 2025-12-04T09:36:41.3494958Z Expected 196608 but got 0. 2025-12-04T09:36:41.3495056Z Absolute difference: 196608 2025-12-04T09:36:41.3495145Z Relative difference: 1.0 2025-12-04T09:36:41.3495148Z 2025-12-04T09:36:41.3495322Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3495718Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3495722Z 2025-12-04T09:36:41.3495940Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3496239Z FAILED [0.1549s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3496243Z 2025-12-04T09:36:41.3496330Z Expected 196608 but got 0. 2025-12-04T09:36:41.3496421Z Absolute difference: 196608 2025-12-04T09:36:41.3496512Z Relative difference: 1.0 2025-12-04T09:36:41.3496516Z 2025-12-04T09:36:41.3496688Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3497076Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3497083Z 2025-12-04T09:36:41.3497295Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3497606Z FAILED [0.1562s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3497610Z 2025-12-04T09:36:41.3497806Z Expected 196608 but got 0. 2025-12-04T09:36:41.3497900Z Absolute difference: 196608 2025-12-04T09:36:41.3497989Z Relative difference: 1.0 2025-12-04T09:36:41.3497993Z 2025-12-04T09:36:41.3498169Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3498558Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3498562Z 2025-12-04T09:36:41.3498777Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3499075Z FAILED [0.1554s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3499079Z 2025-12-04T09:36:41.3499165Z Expected 196608 but got 0. 2025-12-04T09:36:41.3499266Z Absolute difference: 196608 2025-12-04T09:36:41.3499355Z Relative difference: 1.0 2025-12-04T09:36:41.3499359Z 2025-12-04T09:36:41.3499531Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3499926Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3499931Z 2025-12-04T09:36:41.3500146Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3500447Z FAILED [0.1555s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3500451Z 2025-12-04T09:36:41.3500551Z Expected 196608 but got 0. 2025-12-04T09:36:41.3500661Z Absolute difference: 196608 2025-12-04T09:36:41.3500769Z Relative difference: 1.0 2025-12-04T09:36:41.3500774Z 2025-12-04T09:36:41.3500951Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3501343Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3501423Z 2025-12-04T09:36:41.3501636Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3501947Z FAILED [0.1549s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3501951Z 2025-12-04T09:36:41.3502037Z Expected 196608 but got 0. 2025-12-04T09:36:41.3502128Z Absolute difference: 196608 2025-12-04T09:36:41.3502220Z Relative difference: 1.0 2025-12-04T09:36:41.3502224Z 2025-12-04T09:36:41.3502398Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3502786Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3502793Z 2025-12-04T09:36:41.3503007Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3503313Z FAILED [0.1549s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3503317Z 2025-12-04T09:36:41.3503407Z Expected 196608 but got 0. 2025-12-04T09:36:41.3503498Z Absolute difference: 196608 2025-12-04T09:36:41.3503591Z Relative difference: 1.0 2025-12-04T09:36:41.3503595Z 2025-12-04T09:36:41.3503770Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3504158Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3504163Z 2025-12-04T09:36:41.3504377Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3504678Z FAILED [0.1548s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3504682Z 2025-12-04T09:36:41.3504769Z Expected 196608 but got 0. 2025-12-04T09:36:41.3504862Z Absolute difference: 196608 2025-12-04T09:36:41.3504956Z Relative difference: 1.0 2025-12-04T09:36:41.3504960Z 2025-12-04T09:36:41.3505142Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3505607Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3505612Z 2025-12-04T09:36:41.3505825Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3506128Z FAILED [0.1551s] test_cuda.py::TestCuda::test_host_memory_stats - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3506132Z 2025-12-04T09:36:41.3506221Z Expected 196608 but got 0. 2025-12-04T09:36:41.3506313Z Absolute difference: 196608 2025-12-04T09:36:41.3506403Z Relative difference: 1.0 2025-12-04T09:36:41.3506407Z 2025-12-04T09:36:41.3506580Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3506971Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestCuda.test_host_memory_stats 2025-12-04T09:36:41.3506981Z 2025-12-04T09:36:41.3507195Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3507561Z FAILED [0.0017s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3507568Z 2025-12-04T09:36:41.3507655Z Expected 0 but got 123. 2025-12-04T09:36:41.3507745Z Absolute difference: 123 2025-12-04T09:36:41.3507836Z Relative difference: inf 2025-12-04T09:36:41.3507840Z 2025-12-04T09:36:41.3508014Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3508453Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3508458Z 2025-12-04T09:36:41.3508673Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3509032Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3509118Z 2025-12-04T09:36:41.3509209Z Expected 0 but got 123. 2025-12-04T09:36:41.3509299Z Absolute difference: 123 2025-12-04T09:36:41.3509389Z Relative difference: inf 2025-12-04T09:36:41.3509401Z 2025-12-04T09:36:41.3509576Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3510015Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3510020Z 2025-12-04T09:36:41.3510234Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3510601Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3510607Z 2025-12-04T09:36:41.3510710Z Expected 0 but got 123. 2025-12-04T09:36:41.3510814Z Absolute difference: 123 2025-12-04T09:36:41.3510916Z Relative difference: inf 2025-12-04T09:36:41.3510924Z 2025-12-04T09:36:41.3511100Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3511546Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3511551Z 2025-12-04T09:36:41.3511762Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3512120Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3512124Z 2025-12-04T09:36:41.3512210Z Expected 0 but got 123. 2025-12-04T09:36:41.3512301Z Absolute difference: 123 2025-12-04T09:36:41.3512391Z Relative difference: inf 2025-12-04T09:36:41.3512395Z 2025-12-04T09:36:41.3512566Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3513009Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3513018Z 2025-12-04T09:36:41.3513229Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3513667Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3513675Z 2025-12-04T09:36:41.3513762Z Expected 0 but got 123. 2025-12-04T09:36:41.3513851Z Absolute difference: 123 2025-12-04T09:36:41.3513944Z Relative difference: inf 2025-12-04T09:36:41.3513948Z 2025-12-04T09:36:41.3514120Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3514555Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3514560Z 2025-12-04T09:36:41.3514775Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3515132Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3515142Z 2025-12-04T09:36:41.3515231Z Expected 0 but got 123. 2025-12-04T09:36:41.3515321Z Absolute difference: 123 2025-12-04T09:36:41.3515410Z Relative difference: inf 2025-12-04T09:36:41.3515421Z 2025-12-04T09:36:41.3515595Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3516035Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3516040Z 2025-12-04T09:36:41.3516253Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3516612Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3516617Z 2025-12-04T09:36:41.3516702Z Expected 0 but got 123. 2025-12-04T09:36:41.3516794Z Absolute difference: 123 2025-12-04T09:36:41.3516887Z Relative difference: inf 2025-12-04T09:36:41.3516971Z 2025-12-04T09:36:41.3517147Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3517590Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3517595Z 2025-12-04T09:36:41.3517805Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3518162Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3518166Z 2025-12-04T09:36:41.3518252Z Expected 0 but got 123. 2025-12-04T09:36:41.3518342Z Absolute difference: 123 2025-12-04T09:36:41.3518431Z Relative difference: inf 2025-12-04T09:36:41.3518435Z 2025-12-04T09:36:41.3518607Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3519049Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3519061Z 2025-12-04T09:36:41.3519273Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3519631Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3519638Z 2025-12-04T09:36:41.3519726Z Expected 0 but got 123. 2025-12-04T09:36:41.3519815Z Absolute difference: 123 2025-12-04T09:36:41.3519907Z Relative difference: inf 2025-12-04T09:36:41.3519911Z 2025-12-04T09:36:41.3520085Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3520525Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3520529Z 2025-12-04T09:36:41.3520741Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3521096Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3521107Z 2025-12-04T09:36:41.3521196Z Expected 0 but got 123. 2025-12-04T09:36:41.3521286Z Absolute difference: 123 2025-12-04T09:36:41.3521379Z Relative difference: inf 2025-12-04T09:36:41.3521465Z 2025-12-04T09:36:41.3521644Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3522082Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3522087Z 2025-12-04T09:36:41.3522300Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3522656Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3522661Z 2025-12-04T09:36:41.3522748Z Expected 0 but got 123. 2025-12-04T09:36:41.3522841Z Absolute difference: 123 2025-12-04T09:36:41.3522933Z Relative difference: inf 2025-12-04T09:36:41.3522942Z 2025-12-04T09:36:41.3523118Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3523559Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3523564Z 2025-12-04T09:36:41.3523776Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3524134Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3524139Z 2025-12-04T09:36:41.3524224Z Expected 0 but got 123. 2025-12-04T09:36:41.3524315Z Absolute difference: 123 2025-12-04T09:36:41.3524406Z Relative difference: inf 2025-12-04T09:36:41.3524410Z 2025-12-04T09:36:41.3524584Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3525025Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3525149Z 2025-12-04T09:36:41.3525364Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3525726Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3525733Z 2025-12-04T09:36:41.3525820Z Expected 0 but got 123. 2025-12-04T09:36:41.3525910Z Absolute difference: 123 2025-12-04T09:36:41.3526001Z Relative difference: inf 2025-12-04T09:36:41.3526005Z 2025-12-04T09:36:41.3526178Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3526617Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3526621Z 2025-12-04T09:36:41.3526835Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3527194Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3527204Z 2025-12-04T09:36:41.3527343Z Expected 0 but got 123. 2025-12-04T09:36:41.3527434Z Absolute difference: 123 2025-12-04T09:36:41.3527528Z Relative difference: inf 2025-12-04T09:36:41.3527532Z 2025-12-04T09:36:41.3527709Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3528149Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3528154Z 2025-12-04T09:36:41.3528368Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3528723Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3528727Z 2025-12-04T09:36:41.3528813Z Expected 0 but got 123. 2025-12-04T09:36:41.3528907Z Absolute difference: 123 2025-12-04T09:36:41.3528996Z Relative difference: inf 2025-12-04T09:36:41.3529005Z 2025-12-04T09:36:41.3529180Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3529701Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3529706Z 2025-12-04T09:36:41.3529918Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3530276Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3530280Z 2025-12-04T09:36:41.3530366Z Expected 0 but got 123. 2025-12-04T09:36:41.3530455Z Absolute difference: 123 2025-12-04T09:36:41.3530560Z Relative difference: inf 2025-12-04T09:36:41.3530565Z 2025-12-04T09:36:41.3530765Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3531213Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3531221Z 2025-12-04T09:36:41.3531434Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3531798Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3531805Z 2025-12-04T09:36:41.3531896Z Expected 0 but got 123. 2025-12-04T09:36:41.3531985Z Absolute difference: 123 2025-12-04T09:36:41.3532075Z Relative difference: inf 2025-12-04T09:36:41.3532079Z 2025-12-04T09:36:41.3532252Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3532693Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3532697Z 2025-12-04T09:36:41.3532910Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3533267Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3533348Z 2025-12-04T09:36:41.3533437Z Expected 0 but got 123. 2025-12-04T09:36:41.3533531Z Absolute difference: 123 2025-12-04T09:36:41.3533628Z Relative difference: inf 2025-12-04T09:36:41.3533632Z 2025-12-04T09:36:41.3533808Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3534247Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3534251Z 2025-12-04T09:36:41.3534465Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3534819Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3534823Z 2025-12-04T09:36:41.3534908Z Expected 0 but got 123. 2025-12-04T09:36:41.3534999Z Absolute difference: 123 2025-12-04T09:36:41.3535095Z Relative difference: inf 2025-12-04T09:36:41.3535098Z 2025-12-04T09:36:41.3535501Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3535951Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3535955Z 2025-12-04T09:36:41.3536167Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3536529Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3536533Z 2025-12-04T09:36:41.3536621Z Expected 0 but got 123. 2025-12-04T09:36:41.3536712Z Absolute difference: 123 2025-12-04T09:36:41.3536804Z Relative difference: inf 2025-12-04T09:36:41.3536807Z 2025-12-04T09:36:41.3536980Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3537418Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3537428Z 2025-12-04T09:36:41.3537640Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3538205Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3538215Z 2025-12-04T09:36:41.3538302Z Expected 0 but got 123. 2025-12-04T09:36:41.3538391Z Absolute difference: 123 2025-12-04T09:36:41.3538482Z Relative difference: inf 2025-12-04T09:36:41.3538486Z 2025-12-04T09:36:41.3538658Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3539094Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3539098Z 2025-12-04T09:36:41.3539312Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3539666Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3539675Z 2025-12-04T09:36:41.3539764Z Expected 0 but got 123. 2025-12-04T09:36:41.3539852Z Absolute difference: 123 2025-12-04T09:36:41.3539948Z Relative difference: inf 2025-12-04T09:36:41.3539952Z 2025-12-04T09:36:41.3540129Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3540583Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3540588Z 2025-12-04T09:36:41.3540838Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3541193Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3541197Z 2025-12-04T09:36:41.3541282Z Expected 0 but got 123. 2025-12-04T09:36:41.3541376Z Absolute difference: 123 2025-12-04T09:36:41.3541576Z Relative difference: inf 2025-12-04T09:36:41.3541580Z 2025-12-04T09:36:41.3541753Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3542206Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3542210Z 2025-12-04T09:36:41.3542421Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3542776Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3542781Z 2025-12-04T09:36:41.3542870Z Expected 0 but got 123. 2025-12-04T09:36:41.3542959Z Absolute difference: 123 2025-12-04T09:36:41.3543051Z Relative difference: inf 2025-12-04T09:36:41.3543055Z 2025-12-04T09:36:41.3543227Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3543666Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3543675Z 2025-12-04T09:36:41.3543887Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3544245Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3544252Z 2025-12-04T09:36:41.3544338Z Expected 0 but got 123. 2025-12-04T09:36:41.3544426Z Absolute difference: 123 2025-12-04T09:36:41.3544520Z Relative difference: inf 2025-12-04T09:36:41.3544524Z 2025-12-04T09:36:41.3544697Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3545135Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3545139Z 2025-12-04T09:36:41.3545352Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3545706Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3545716Z 2025-12-04T09:36:41.3545805Z Expected 0 but got 123. 2025-12-04T09:36:41.3545894Z Absolute difference: 123 2025-12-04T09:36:41.3546064Z Relative difference: inf 2025-12-04T09:36:41.3546069Z 2025-12-04T09:36:41.3546245Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3546681Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3546685Z 2025-12-04T09:36:41.3546899Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3547252Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3547256Z 2025-12-04T09:36:41.3547342Z Expected 0 but got 123. 2025-12-04T09:36:41.3547436Z Absolute difference: 123 2025-12-04T09:36:41.3547530Z Relative difference: inf 2025-12-04T09:36:41.3547534Z 2025-12-04T09:36:41.3547707Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3548157Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3548161Z 2025-12-04T09:36:41.3548372Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3548733Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3548737Z 2025-12-04T09:36:41.3548823Z Expected 0 but got 123. 2025-12-04T09:36:41.3548913Z Absolute difference: 123 2025-12-04T09:36:41.3549004Z Relative difference: inf 2025-12-04T09:36:41.3549008Z 2025-12-04T09:36:41.3549182Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3549622Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3549706Z 2025-12-04T09:36:41.3549919Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3550281Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3550288Z 2025-12-04T09:36:41.3550374Z Expected 0 but got 123. 2025-12-04T09:36:41.3550468Z Absolute difference: 123 2025-12-04T09:36:41.3550559Z Relative difference: inf 2025-12-04T09:36:41.3550563Z 2025-12-04T09:36:41.3550737Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3551178Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3551183Z 2025-12-04T09:36:41.3551403Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3551767Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3551771Z 2025-12-04T09:36:41.3551859Z Expected 0 but got 123. 2025-12-04T09:36:41.3551948Z Absolute difference: 123 2025-12-04T09:36:41.3552039Z Relative difference: inf 2025-12-04T09:36:41.3552044Z 2025-12-04T09:36:41.3552219Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3552661Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3552665Z 2025-12-04T09:36:41.3552879Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3553234Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3553238Z 2025-12-04T09:36:41.3553324Z Expected 0 but got 123. 2025-12-04T09:36:41.3553417Z Absolute difference: 123 2025-12-04T09:36:41.3553511Z Relative difference: inf 2025-12-04T09:36:41.3553515Z 2025-12-04T09:36:41.3553687Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3554207Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3554211Z 2025-12-04T09:36:41.3554424Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3554781Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3554786Z 2025-12-04T09:36:41.3554871Z Expected 0 but got 123. 2025-12-04T09:36:41.3554960Z Absolute difference: 123 2025-12-04T09:36:41.3555050Z Relative difference: inf 2025-12-04T09:36:41.3555054Z 2025-12-04T09:36:41.3555227Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3555666Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3555675Z 2025-12-04T09:36:41.3555888Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3556247Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3556255Z 2025-12-04T09:36:41.3556341Z Expected 0 but got 123. 2025-12-04T09:36:41.3556430Z Absolute difference: 123 2025-12-04T09:36:41.3556522Z Relative difference: inf 2025-12-04T09:36:41.3556526Z 2025-12-04T09:36:41.3556699Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3557138Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3557143Z 2025-12-04T09:36:41.3557357Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3557817Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3557821Z 2025-12-04T09:36:41.3557910Z Expected 0 but got 123. 2025-12-04T09:36:41.3557999Z Absolute difference: 123 2025-12-04T09:36:41.3558093Z Relative difference: inf 2025-12-04T09:36:41.3558097Z 2025-12-04T09:36:41.3558273Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3558714Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3558719Z 2025-12-04T09:36:41.3558935Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3559290Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3559294Z 2025-12-04T09:36:41.3559380Z Expected 0 but got 123. 2025-12-04T09:36:41.3559473Z Absolute difference: 123 2025-12-04T09:36:41.3559569Z Relative difference: inf 2025-12-04T09:36:41.3559573Z 2025-12-04T09:36:41.3559747Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3560192Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3560196Z 2025-12-04T09:36:41.3560409Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3560772Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3560778Z 2025-12-04T09:36:41.3560886Z Expected 0 but got 123. 2025-12-04T09:36:41.3560982Z Absolute difference: 123 2025-12-04T09:36:41.3561091Z Relative difference: inf 2025-12-04T09:36:41.3561095Z 2025-12-04T09:36:41.3561267Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3561708Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3561718Z 2025-12-04T09:36:41.3561932Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3562363Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3562372Z 2025-12-04T09:36:41.3562460Z Expected 0 but got 123. 2025-12-04T09:36:41.3562549Z Absolute difference: 123 2025-12-04T09:36:41.3562648Z Relative difference: inf 2025-12-04T09:36:41.3562652Z 2025-12-04T09:36:41.3562825Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3563264Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3563269Z 2025-12-04T09:36:41.3563482Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3563843Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3563847Z 2025-12-04T09:36:41.3563937Z Expected 0 but got 123. 2025-12-04T09:36:41.3564034Z Absolute difference: 123 2025-12-04T09:36:41.3564123Z Relative difference: inf 2025-12-04T09:36:41.3564127Z 2025-12-04T09:36:41.3564303Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3564742Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3564747Z 2025-12-04T09:36:41.3564961Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3565317Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3565321Z 2025-12-04T09:36:41.3565407Z Expected 0 but got 123. 2025-12-04T09:36:41.3565499Z Absolute difference: 123 2025-12-04T09:36:41.3565667Z Relative difference: inf 2025-12-04T09:36:41.3565671Z 2025-12-04T09:36:41.3565844Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3566289Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3566294Z 2025-12-04T09:36:41.3566505Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3566862Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3566867Z 2025-12-04T09:36:41.3566953Z Expected 0 but got 123. 2025-12-04T09:36:41.3567041Z Absolute difference: 123 2025-12-04T09:36:41.3567134Z Relative difference: inf 2025-12-04T09:36:41.3567138Z 2025-12-04T09:36:41.3567357Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3567797Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3567810Z 2025-12-04T09:36:41.3568023Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3568382Z FAILED [0.0019s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3568386Z 2025-12-04T09:36:41.3568475Z Expected 0 but got 123. 2025-12-04T09:36:41.3568564Z Absolute difference: 123 2025-12-04T09:36:41.3568655Z Relative difference: inf 2025-12-04T09:36:41.3568660Z 2025-12-04T09:36:41.3568831Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3569272Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3569276Z 2025-12-04T09:36:41.3569492Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3569855Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3569860Z 2025-12-04T09:36:41.3569949Z Expected 0 but got 123. 2025-12-04T09:36:41.3570116Z Absolute difference: 123 2025-12-04T09:36:41.3570207Z Relative difference: inf 2025-12-04T09:36:41.3570211Z 2025-12-04T09:36:41.3570387Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3570828Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3570832Z 2025-12-04T09:36:41.3571047Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3571404Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3571408Z 2025-12-04T09:36:41.3571496Z Expected 0 but got 123. 2025-12-04T09:36:41.3571591Z Absolute difference: 123 2025-12-04T09:36:41.3571679Z Relative difference: inf 2025-12-04T09:36:41.3571683Z 2025-12-04T09:36:41.3571855Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3572298Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3572302Z 2025-12-04T09:36:41.3572514Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3572874Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3572878Z 2025-12-04T09:36:41.3572967Z Expected 0 but got 123. 2025-12-04T09:36:41.3573057Z Absolute difference: 123 2025-12-04T09:36:41.3573149Z Relative difference: inf 2025-12-04T09:36:41.3573152Z 2025-12-04T09:36:41.3573324Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3573762Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3573844Z 2025-12-04T09:36:41.3574059Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3574422Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3574426Z 2025-12-04T09:36:41.3574516Z Expected 0 but got 123. 2025-12-04T09:36:41.3574604Z Absolute difference: 123 2025-12-04T09:36:41.3574695Z Relative difference: inf 2025-12-04T09:36:41.3574699Z 2025-12-04T09:36:41.3574871Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3575312Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3575316Z 2025-12-04T09:36:41.3575535Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3575899Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3575903Z 2025-12-04T09:36:41.3575993Z Expected 0 but got 123. 2025-12-04T09:36:41.3576085Z Absolute difference: 123 2025-12-04T09:36:41.3576173Z Relative difference: inf 2025-12-04T09:36:41.3576177Z 2025-12-04T09:36:41.3576353Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3576793Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3576797Z 2025-12-04T09:36:41.3577011Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3577370Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3577374Z 2025-12-04T09:36:41.3577460Z Expected 0 but got 123. 2025-12-04T09:36:41.3577555Z Absolute difference: 123 2025-12-04T09:36:41.3577642Z Relative difference: inf 2025-12-04T09:36:41.3577646Z 2025-12-04T09:36:41.3577819Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3578346Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3578351Z 2025-12-04T09:36:41.3578565Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3578923Z FAILED [0.0013s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3578927Z 2025-12-04T09:36:41.3579013Z Expected 0 but got 123. 2025-12-04T09:36:41.3579102Z Absolute difference: 123 2025-12-04T09:36:41.3579194Z Relative difference: inf 2025-12-04T09:36:41.3579198Z 2025-12-04T09:36:41.3579372Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3583090Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3583105Z 2025-12-04T09:36:41.3583341Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3583711Z FAILED [0.0014s] test_cuda.py::TestMemPool::test_mempool_empty_cache_inactive - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3583716Z 2025-12-04T09:36:41.3583809Z Expected 0 but got 123. 2025-12-04T09:36:41.3583901Z Absolute difference: 123 2025-12-04T09:36:41.3583994Z Relative difference: inf 2025-12-04T09:36:41.3583998Z 2025-12-04T09:36:41.3584174Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3584620Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_empty_cache_inactive 2025-12-04T09:36:41.3584624Z 2025-12-04T09:36:41.3584840Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3585288Z FAILED [0.0019s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3585293Z 2025-12-04T09:36:41.3585382Z Expected 0 but got 123. 2025-12-04T09:36:41.3585477Z Absolute difference: 123 2025-12-04T09:36:41.3585567Z Relative difference: inf 2025-12-04T09:36:41.3585571Z 2025-12-04T09:36:41.3585752Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3586176Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3586181Z 2025-12-04T09:36:41.3586392Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3586730Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3586734Z 2025-12-04T09:36:41.3586822Z Expected 0 but got 123. 2025-12-04T09:36:41.3586915Z Absolute difference: 123 2025-12-04T09:36:41.3587009Z Relative difference: inf 2025-12-04T09:36:41.3587013Z 2025-12-04T09:36:41.3587189Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3587618Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3587623Z 2025-12-04T09:36:41.3587835Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3588178Z FAILED [0.0019s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3588183Z 2025-12-04T09:36:41.3588271Z Expected 0 but got 123. 2025-12-04T09:36:41.3588361Z Absolute difference: 123 2025-12-04T09:36:41.3588454Z Relative difference: inf 2025-12-04T09:36:41.3588458Z 2025-12-04T09:36:41.3588631Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3589054Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3589063Z 2025-12-04T09:36:41.3589274Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3589712Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3589717Z 2025-12-04T09:36:41.3589807Z Expected 0 but got 123. 2025-12-04T09:36:41.3589896Z Absolute difference: 123 2025-12-04T09:36:41.3589987Z Relative difference: inf 2025-12-04T09:36:41.3589994Z 2025-12-04T09:36:41.3590167Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3590586Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3590591Z 2025-12-04T09:36:41.3590805Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3591153Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3591157Z 2025-12-04T09:36:41.3591246Z Expected 0 but got 123. 2025-12-04T09:36:41.3591336Z Absolute difference: 123 2025-12-04T09:36:41.3591430Z Relative difference: inf 2025-12-04T09:36:41.3591435Z 2025-12-04T09:36:41.3591610Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3592031Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3592035Z 2025-12-04T09:36:41.3592246Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3592592Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3592596Z 2025-12-04T09:36:41.3592682Z Expected 0 but got 123. 2025-12-04T09:36:41.3592773Z Absolute difference: 123 2025-12-04T09:36:41.3592939Z Relative difference: inf 2025-12-04T09:36:41.3592943Z 2025-12-04T09:36:41.3593121Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3593551Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3593555Z 2025-12-04T09:36:41.3593766Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3594105Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3594109Z 2025-12-04T09:36:41.3594194Z Expected 0 but got 123. 2025-12-04T09:36:41.3594284Z Absolute difference: 123 2025-12-04T09:36:41.3594376Z Relative difference: inf 2025-12-04T09:36:41.3594380Z 2025-12-04T09:36:41.3594553Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3594978Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3594988Z 2025-12-04T09:36:41.3595201Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3595540Z FAILED [0.0018s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3595544Z 2025-12-04T09:36:41.3595633Z Expected 0 but got 123. 2025-12-04T09:36:41.3595722Z Absolute difference: 123 2025-12-04T09:36:41.3595811Z Relative difference: inf 2025-12-04T09:36:41.3595818Z 2025-12-04T09:36:41.3595990Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3596412Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3596416Z 2025-12-04T09:36:41.3596630Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3596964Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3596976Z 2025-12-04T09:36:41.3597062Z Expected 0 but got 123. 2025-12-04T09:36:41.3597160Z Absolute difference: 123 2025-12-04T09:36:41.3597249Z Relative difference: inf 2025-12-04T09:36:41.3597331Z 2025-12-04T09:36:41.3597509Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3597929Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3597933Z 2025-12-04T09:36:41.3598143Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3598479Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3598484Z 2025-12-04T09:36:41.3598570Z Expected 0 but got 123. 2025-12-04T09:36:41.3598662Z Absolute difference: 123 2025-12-04T09:36:41.3598752Z Relative difference: inf 2025-12-04T09:36:41.3598764Z 2025-12-04T09:36:41.3598937Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3599367Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3599371Z 2025-12-04T09:36:41.3599582Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3599922Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3599926Z 2025-12-04T09:36:41.3600010Z Expected 0 but got 123. 2025-12-04T09:36:41.3600098Z Absolute difference: 123 2025-12-04T09:36:41.3600189Z Relative difference: inf 2025-12-04T09:36:41.3600193Z 2025-12-04T09:36:41.3600366Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3600786Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3600870Z 2025-12-04T09:36:41.3601082Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3601432Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3601438Z 2025-12-04T09:36:41.3601546Z Expected 0 but got 123. 2025-12-04T09:36:41.3601644Z Absolute difference: 123 2025-12-04T09:36:41.3601752Z Relative difference: inf 2025-12-04T09:36:41.3601756Z 2025-12-04T09:36:41.3601931Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3602353Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3602357Z 2025-12-04T09:36:41.3602570Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3602905Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3602914Z 2025-12-04T09:36:41.3603000Z Expected 0 but got 123. 2025-12-04T09:36:41.3603094Z Absolute difference: 123 2025-12-04T09:36:41.3603183Z Relative difference: inf 2025-12-04T09:36:41.3603187Z 2025-12-04T09:36:41.3603369Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3603792Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3603796Z 2025-12-04T09:36:41.3604007Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3604344Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3604349Z 2025-12-04T09:36:41.3604434Z Expected 0 but got 123. 2025-12-04T09:36:41.3604531Z Absolute difference: 123 2025-12-04T09:36:41.3604618Z Relative difference: inf 2025-12-04T09:36:41.3604622Z 2025-12-04T09:36:41.3604799Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3605222Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3605302Z 2025-12-04T09:36:41.3605518Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3605860Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3605864Z 2025-12-04T09:36:41.3605951Z Expected 0 but got 123. 2025-12-04T09:36:41.3606042Z Absolute difference: 123 2025-12-04T09:36:41.3606133Z Relative difference: inf 2025-12-04T09:36:41.3606137Z 2025-12-04T09:36:41.3606310Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3606732Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3606745Z 2025-12-04T09:36:41.3606957Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3607354Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3607363Z 2025-12-04T09:36:41.3607453Z Expected 0 but got 123. 2025-12-04T09:36:41.3607542Z Absolute difference: 123 2025-12-04T09:36:41.3607631Z Relative difference: inf 2025-12-04T09:36:41.3607635Z 2025-12-04T09:36:41.3607812Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3608232Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3608236Z 2025-12-04T09:36:41.3608449Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3608782Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3608868Z 2025-12-04T09:36:41.3608956Z Expected 0 but got 123. 2025-12-04T09:36:41.3609048Z Absolute difference: 123 2025-12-04T09:36:41.3609136Z Relative difference: inf 2025-12-04T09:36:41.3609140Z 2025-12-04T09:36:41.3609320Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3609739Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3609743Z 2025-12-04T09:36:41.3609954Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3610293Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3610297Z 2025-12-04T09:36:41.3610382Z Expected 0 but got 123. 2025-12-04T09:36:41.3610475Z Absolute difference: 123 2025-12-04T09:36:41.3610568Z Relative difference: inf 2025-12-04T09:36:41.3610571Z 2025-12-04T09:36:41.3610747Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3611225Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3611229Z 2025-12-04T09:36:41.3611447Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3611782Z FAILED [0.0017s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3611790Z 2025-12-04T09:36:41.3611875Z Expected 0 but got 123. 2025-12-04T09:36:41.3611965Z Absolute difference: 123 2025-12-04T09:36:41.3612059Z Relative difference: inf 2025-12-04T09:36:41.3612063Z 2025-12-04T09:36:41.3612237Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3612657Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3612662Z 2025-12-04T09:36:41.3612884Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3613218Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3613222Z 2025-12-04T09:36:41.3613388Z Expected 0 but got 123. 2025-12-04T09:36:41.3613479Z Absolute difference: 123 2025-12-04T09:36:41.3613568Z Relative difference: inf 2025-12-04T09:36:41.3613573Z 2025-12-04T09:36:41.3613747Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3614167Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3614171Z 2025-12-04T09:36:41.3614385Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3614721Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3614725Z 2025-12-04T09:36:41.3614815Z Expected 0 but got 123. 2025-12-04T09:36:41.3614909Z Absolute difference: 123 2025-12-04T09:36:41.3614998Z Relative difference: inf 2025-12-04T09:36:41.3615002Z 2025-12-04T09:36:41.3615176Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3615599Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3615603Z 2025-12-04T09:36:41.3615814Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3616150Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3616154Z 2025-12-04T09:36:41.3616241Z Expected 0 but got 123. 2025-12-04T09:36:41.3616329Z Absolute difference: 123 2025-12-04T09:36:41.3616420Z Relative difference: inf 2025-12-04T09:36:41.3616424Z 2025-12-04T09:36:41.3616597Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3617095Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3617099Z 2025-12-04T09:36:41.3617315Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3617648Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3617655Z 2025-12-04T09:36:41.3617740Z Expected 0 but got 123. 2025-12-04T09:36:41.3617829Z Absolute difference: 123 2025-12-04T09:36:41.3617922Z Relative difference: inf 2025-12-04T09:36:41.3617926Z 2025-12-04T09:36:41.3618099Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3618518Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3618523Z 2025-12-04T09:36:41.3618736Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3619076Z FAILED [0.0017s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3619081Z 2025-12-04T09:36:41.3619173Z Expected 0 but got 123. 2025-12-04T09:36:41.3619266Z Absolute difference: 123 2025-12-04T09:36:41.3619355Z Relative difference: inf 2025-12-04T09:36:41.3619360Z 2025-12-04T09:36:41.3619537Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3619956Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3619960Z 2025-12-04T09:36:41.3620174Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3620525Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3620530Z 2025-12-04T09:36:41.3620625Z Expected 0 but got 123. 2025-12-04T09:36:41.3620743Z Absolute difference: 123 2025-12-04T09:36:41.3620835Z Relative difference: inf 2025-12-04T09:36:41.3620839Z 2025-12-04T09:36:41.3621013Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3621567Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3621571Z 2025-12-04T09:36:41.3621787Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3622124Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3622128Z 2025-12-04T09:36:41.3622215Z Expected 0 but got 123. 2025-12-04T09:36:41.3622304Z Absolute difference: 123 2025-12-04T09:36:41.3622397Z Relative difference: inf 2025-12-04T09:36:41.3622401Z 2025-12-04T09:36:41.3622573Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3622996Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3623008Z 2025-12-04T09:36:41.3623222Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3623560Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3623568Z 2025-12-04T09:36:41.3623655Z Expected 0 but got 123. 2025-12-04T09:36:41.3623744Z Absolute difference: 123 2025-12-04T09:36:41.3623836Z Relative difference: inf 2025-12-04T09:36:41.3623840Z 2025-12-04T09:36:41.3624014Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3624435Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3624440Z 2025-12-04T09:36:41.3624654Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3625068Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3625073Z 2025-12-04T09:36:41.3625162Z Expected 0 but got 123. 2025-12-04T09:36:41.3625257Z Absolute difference: 123 2025-12-04T09:36:41.3625346Z Relative difference: inf 2025-12-04T09:36:41.3625351Z 2025-12-04T09:36:41.3625527Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3625948Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3625953Z 2025-12-04T09:36:41.3626166Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3626503Z FAILED [0.0017s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3626507Z 2025-12-04T09:36:41.3626592Z Expected 0 but got 123. 2025-12-04T09:36:41.3626688Z Absolute difference: 123 2025-12-04T09:36:41.3626783Z Relative difference: inf 2025-12-04T09:36:41.3626787Z 2025-12-04T09:36:41.3626961Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3627388Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3627392Z 2025-12-04T09:36:41.3627604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3627941Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3627945Z 2025-12-04T09:36:41.3628031Z Expected 0 but got 123. 2025-12-04T09:36:41.3628119Z Absolute difference: 123 2025-12-04T09:36:41.3628211Z Relative difference: inf 2025-12-04T09:36:41.3628215Z 2025-12-04T09:36:41.3628390Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3628810Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3628819Z 2025-12-04T09:36:41.3629029Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3629446Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3629451Z 2025-12-04T09:36:41.3629540Z Expected 0 but got 123. 2025-12-04T09:36:41.3629628Z Absolute difference: 123 2025-12-04T09:36:41.3629718Z Relative difference: inf 2025-12-04T09:36:41.3629722Z 2025-12-04T09:36:41.3629896Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3630318Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3630322Z 2025-12-04T09:36:41.3630535Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3630895Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3630905Z 2025-12-04T09:36:41.3631012Z Expected 0 but got 123. 2025-12-04T09:36:41.3631108Z Absolute difference: 123 2025-12-04T09:36:41.3631200Z Relative difference: inf 2025-12-04T09:36:41.3631204Z 2025-12-04T09:36:41.3631380Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3631801Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3631805Z 2025-12-04T09:36:41.3632020Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3632355Z FAILED [0.0018s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3632359Z 2025-12-04T09:36:41.3632445Z Expected 0 but got 123. 2025-12-04T09:36:41.3632537Z Absolute difference: 123 2025-12-04T09:36:41.3632627Z Relative difference: inf 2025-12-04T09:36:41.3632709Z 2025-12-04T09:36:41.3632883Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3633309Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3633314Z 2025-12-04T09:36:41.3633525Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3633863Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3633867Z 2025-12-04T09:36:41.3633952Z Expected 0 but got 123. 2025-12-04T09:36:41.3634041Z Absolute difference: 123 2025-12-04T09:36:41.3634133Z Relative difference: inf 2025-12-04T09:36:41.3634137Z 2025-12-04T09:36:41.3634309Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3634731Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3634742Z 2025-12-04T09:36:41.3634953Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3635460Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3635464Z 2025-12-04T09:36:41.3635556Z Expected 0 but got 123. 2025-12-04T09:36:41.3635646Z Absolute difference: 123 2025-12-04T09:36:41.3635738Z Relative difference: inf 2025-12-04T09:36:41.3635742Z 2025-12-04T09:36:41.3635919Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3636340Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3636344Z 2025-12-04T09:36:41.3636560Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3636897Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3636910Z 2025-12-04T09:36:41.3636999Z Expected 0 but got 123. 2025-12-04T09:36:41.3637087Z Absolute difference: 123 2025-12-04T09:36:41.3637176Z Relative difference: inf 2025-12-04T09:36:41.3637302Z 2025-12-04T09:36:41.3637484Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3637905Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3637909Z 2025-12-04T09:36:41.3638120Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3638458Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3638463Z 2025-12-04T09:36:41.3638550Z Expected 0 but got 123. 2025-12-04T09:36:41.3638642Z Absolute difference: 123 2025-12-04T09:36:41.3638731Z Relative difference: inf 2025-12-04T09:36:41.3638739Z 2025-12-04T09:36:41.3638912Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3639341Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3639345Z 2025-12-04T09:36:41.3639556Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3639893Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3639897Z 2025-12-04T09:36:41.3639984Z Expected 0 but got 123. 2025-12-04T09:36:41.3640073Z Absolute difference: 123 2025-12-04T09:36:41.3640164Z Relative difference: inf 2025-12-04T09:36:41.3640168Z 2025-12-04T09:36:41.3640341Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3640764Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3640876Z 2025-12-04T09:36:41.3641089Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3641434Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3641438Z 2025-12-04T09:36:41.3641529Z Expected 0 but got 123. 2025-12-04T09:36:41.3641618Z Absolute difference: 123 2025-12-04T09:36:41.3641709Z Relative difference: inf 2025-12-04T09:36:41.3641715Z 2025-12-04T09:36:41.3641890Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3642311Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3642315Z 2025-12-04T09:36:41.3642530Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3642865Z FAILED [0.0018s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3642874Z 2025-12-04T09:36:41.3642963Z Expected 0 but got 123. 2025-12-04T09:36:41.3643053Z Absolute difference: 123 2025-12-04T09:36:41.3643143Z Relative difference: inf 2025-12-04T09:36:41.3643147Z 2025-12-04T09:36:41.3643327Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3643748Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3643752Z 2025-12-04T09:36:41.3643964Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3644304Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3644308Z 2025-12-04T09:36:41.3644395Z Expected 0 but got 123. 2025-12-04T09:36:41.3644489Z Absolute difference: 123 2025-12-04T09:36:41.3644578Z Relative difference: inf 2025-12-04T09:36:41.3644582Z 2025-12-04T09:36:41.3644761Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3645182Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3645186Z 2025-12-04T09:36:41.3645481Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3645821Z FAILED [0.0016s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3645825Z 2025-12-04T09:36:41.3645910Z Expected 0 but got 123. 2025-12-04T09:36:41.3646000Z Absolute difference: 123 2025-12-04T09:36:41.3646092Z Relative difference: inf 2025-12-04T09:36:41.3646096Z 2025-12-04T09:36:41.3646273Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3646693Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3646700Z 2025-12-04T09:36:41.3646918Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3647256Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3647312Z 2025-12-04T09:36:41.3647401Z Expected 0 but got 123. 2025-12-04T09:36:41.3647489Z Absolute difference: 123 2025-12-04T09:36:41.3647577Z Relative difference: inf 2025-12-04T09:36:41.3647582Z 2025-12-04T09:36:41.3647763Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3648183Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3648187Z 2025-12-04T09:36:41.3648402Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3648735Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3648844Z 2025-12-04T09:36:41.3648931Z Expected 0 but got 123. 2025-12-04T09:36:41.3649023Z Absolute difference: 123 2025-12-04T09:36:41.3649111Z Relative difference: inf 2025-12-04T09:36:41.3649116Z 2025-12-04T09:36:41.3649301Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3649722Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3649727Z 2025-12-04T09:36:41.3649938Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3650278Z FAILED [0.0017s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3650282Z 2025-12-04T09:36:41.3650367Z Expected 0 but got 123. 2025-12-04T09:36:41.3650458Z Absolute difference: 123 2025-12-04T09:36:41.3650555Z Relative difference: inf 2025-12-04T09:36:41.3650560Z 2025-12-04T09:36:41.3650760Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3651207Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3651211Z 2025-12-04T09:36:41.3651424Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3651763Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3651767Z 2025-12-04T09:36:41.3651853Z Expected 0 but got 123. 2025-12-04T09:36:41.3651942Z Absolute difference: 123 2025-12-04T09:36:41.3652034Z Relative difference: inf 2025-12-04T09:36:41.3652038Z 2025-12-04T09:36:41.3652211Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3652634Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3652641Z 2025-12-04T09:36:41.3652852Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3653193Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3653197Z 2025-12-04T09:36:41.3653372Z Expected 0 but got 123. 2025-12-04T09:36:41.3653468Z Absolute difference: 123 2025-12-04T09:36:41.3653556Z Relative difference: inf 2025-12-04T09:36:41.3653560Z 2025-12-04T09:36:41.3653737Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3654162Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3654166Z 2025-12-04T09:36:41.3654386Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3654724Z FAILED [0.0015s] test_cuda.py::TestMemPool::test_mempool_with_allocator - AssertionError: Scalars are not equal! 2025-12-04T09:36:41.3654728Z 2025-12-04T09:36:41.3654814Z Expected 0 but got 123. 2025-12-04T09:36:41.3654913Z Absolute difference: 123 2025-12-04T09:36:41.3655001Z Relative difference: inf 2025-12-04T09:36:41.3655006Z 2025-12-04T09:36:41.3655182Z To execute this test, run the following from the base repo dir: 2025-12-04T09:36:41.3655609Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda.py TestMemPool.test_mempool_with_allocator 2025-12-04T09:36:41.3655613Z 2025-12-04T09:36:41.3655823Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:36:41.3656001Z =========== 149 failed, 251 passed, 150 skipped in 104.95s (0:01:44) =========== 2025-12-04T09:36:41.3656006Z 2025-12-04T09:36:41.3656294Z FINISHED PRINTING LOG FILE of test_cuda 1/1 (test/test-reports/test_cuda_1.1_2d01c4f5126b94de_.log) 2025-12-04T09:36:41.3656299Z 2025-12-04T09:36:41.3656500Z Finished test_cuda 1/1 ... [2025-12-04 09:36:41.177558][1503.594027673], took 1.84min 2025-12-04T09:36:41.3656978Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda/test_cuda-be50f111eb28e9fa.xml 2025-12-04T09:36:41.5013864Z Uploading logs for 57118183167 to S3 2025-12-04T09:36:41.5524049Z Uploading artifacts took 0.28 seconds 2025-12-04T09:36:41.5524360Z test_cuda 1/1 failed! 2025-12-04T09:36:41.5531284Z Running test_sparse 1/1 ... [2025-12-04 09:36:41.552560][1503.96903132] 2025-12-04T09:36:41.5531691Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:36:41.5533113Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_sparse.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:36:41.552936] 2025-12-04T09:36:51.6866844Z 2025-12-04T09:36:51.6868809Z test_sparse 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_sparse_1.1_f9df9789c45a0117_.log 2025-12-04T09:36:51.6870082Z Running 0 items in this shard: 2025-12-04T09:36:51.6870498Z 2025-12-04T09:36:51.6870943Z Finished test_sparse 1/1 ... [2025-12-04 09:36:51.686281][1514.102753507], took 0.17min 2025-12-04T09:36:51.6899120Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_sparse/test_sparse-1f8fd945d58b94e8.xml 2025-12-04T09:36:51.7630731Z Running test_ci_sanity_check_fail 1/1 ... [2025-12-04 09:36:51.762676][1514.179148522] 2025-12-04T09:36:51.7631183Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:36:51.7634187Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ci_sanity_check_fail.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:36:51.763031] 2025-12-04T09:36:54.9837804Z Finished test_ci_sanity_check_fail 1/1 ... [2025-12-04 09:36:54.983116][1517.399575943], took 0.05min 2025-12-04T09:36:54.9871571Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ci_sanity_check_fail/test_ci_sanity_check_fail-024947735ba5c526.xml 2025-12-04T09:36:55.2987425Z Uploading logs for 57118183167 to S3 2025-12-04T09:36:55.3302091Z Uploading artifacts took 0.32 seconds 2025-12-04T09:36:55.3302423Z test_ci_sanity_check_fail 1/1 failed! 2025-12-04T09:36:55.3306372Z Running test_ops_fwd_gradients 6/12 ... [2025-12-04 09:36:55.330326][1517.746798419] 2025-12-04T09:36:55.3306846Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:36:55.3310750Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_fwd_gradients.py', '--shard-id=6', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:36:55.330692] 2025-12-04T09:37:05.3638445Z 2025-12-04T09:37:05.3639377Z test_ops_fwd_gradients 6/12 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_fwd_gradients_6.12_48d002fd493d49ca_.log 2025-12-04T09:37:05.3640137Z Running 0 items in this shard: 2025-12-04T09:37:05.3640324Z 2025-12-04T09:37:05.3640602Z Finished test_ops_fwd_gradients 6/12 ... [2025-12-04 09:37:05.363275][1527.779744896], took 0.17min 2025-12-04T09:37:05.3674923Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-bce2e111090231c8.xml 2025-12-04T09:37:05.4431894Z Running test_ops_gradients 2/10 ... [2025-12-04 09:37:05.442829][1527.859301377] 2025-12-04T09:37:05.4432347Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:37:05.4436152Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_gradients.py', '--shard-id=2', '--num-shards=10', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:37:05.443234] 2025-12-04T09:37:19.2311340Z 2025-12-04T09:37:19.2312234Z test_ops_gradients 2/10 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_gradients_2.10_5a1c7cb33b683fe9_.log 2025-12-04T09:37:19.2312963Z Running 0 items in this shard: 2025-12-04T09:37:19.2313164Z 2025-12-04T09:37:19.2313415Z Finished test_ops_gradients 2/10 ... [2025-12-04 09:37:19.230754][1541.647227759], took 0.23min 2025-12-04T09:37:19.2349375Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_gradients/test_ops_gradients-d325fd253497d5cc.xml 2025-12-04T09:37:19.2963107Z Running test_ops_gradients 10/10 ... [2025-12-04 09:37:19.295874][1541.712347224] 2025-12-04T09:37:19.2963565Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:37:19.2966328Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_gradients.py', '--shard-id=10', '--num-shards=10', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:37:19.296193] 2025-12-04T09:37:33.1353463Z 2025-12-04T09:37:33.1354527Z test_ops_gradients 10/10 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_gradients_10.10_499b989a7051c9dd_.log 2025-12-04T09:37:33.1355244Z Running 0 items in this shard: 2025-12-04T09:37:33.1355430Z 2025-12-04T09:37:33.1355692Z Finished test_ops_gradients 10/10 ... [2025-12-04 09:37:33.135032][1555.551504898], took 0.23min 2025-12-04T09:37:33.1392556Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_gradients/test_ops_gradients-192c0166610ef1c8.xml 2025-12-04T09:37:33.2179225Z Running functorch/test_ops 3/6 ... [2025-12-04 09:37:33.217627][1555.634099927] 2025-12-04T09:37:33.2179762Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:37:33.2184431Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ops.py', '--shard-id=3', '--num-shards=6', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:37:33.217952] 2025-12-04T09:37:55.0722916Z 2025-12-04T09:37:55.0724309Z functorch/test_ops 3/6 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_3.6_223700a0811fc5e7_.log 2025-12-04T09:37:55.0725007Z Running 0 items in this shard: 2025-12-04T09:37:55.0725203Z 2025-12-04T09:37:55.0725457Z Finished functorch/test_ops 3/6 ... [2025-12-04 09:37:55.071897][1577.488369756], took 0.36min 2025-12-04T09:37:55.0763955Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-423c2c6c6eabe2c7.xml 2025-12-04T09:37:55.1495665Z Running dynamo/test_after_aot 1/1 ... [2025-12-04 09:37:55.149183][1577.565655732] 2025-12-04T09:37:55.1496176Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:37:55.1499288Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_after_aot.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:37:55.149539] 2025-12-04T09:37:58.6701647Z 2025-12-04T09:37:58.6702598Z dynamo/test_after_aot 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_after_aot_1.1_27387dde59e614f4_.log 2025-12-04T09:37:58.6703520Z Running 0 items in this shard: 2025-12-04T09:37:58.6703705Z 2025-12-04T09:37:58.6703971Z Finished dynamo/test_after_aot 1/1 ... [2025-12-04 09:37:58.669884][1581.086356968], took 0.06min 2025-12-04T09:37:58.6747288Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_after_aot/dynamo.test_after_aot-a91b437d94dc7298.xml 2025-12-04T09:37:58.7005541Z Running inductor/test_snode_runtime 1/1 ... [2025-12-04 09:37:58.700212][1581.116685192] 2025-12-04T09:37:58.7006587Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:37:58.7009169Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_snode_runtime.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:37:58.700536] 2025-12-04T09:38:04.9257197Z 2025-12-04T09:38:04.9258301Z inductor/test_snode_runtime 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_snode_runtime_1.1_c167dbf323fadd03_.log 2025-12-04T09:38:04.9259071Z Running 0 items in this shard: 2025-12-04T09:38:04.9259265Z 2025-12-04T09:38:04.9259552Z Finished inductor/test_snode_runtime 1/1 ... [2025-12-04 09:38:04.925386][1587.34186022], took 0.10min 2025-12-04T09:38:04.9302128Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_snode_runtime/inductor.test_snode_runtime-1e765d10347efbf7.xml 2025-12-04T09:38:04.9850952Z Running inductor/test_compiled_autograd 1/2 ... [2025-12-04 09:38:04.984711][1587.401183546] 2025-12-04T09:38:04.9851452Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:38:04.9854554Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_compiled_autograd.py', '--shard-id=1', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:38:04.985059] 2025-12-04T09:38:23.0844882Z 2025-12-04T09:38:23.0845587Z PRINTING LOG FILE of inductor/test_compiled_autograd 1/2 (test/test-reports/inductor.test_compiled_autograd_1.2_1749786211dd0efc_.log) 2025-12-04T09:38:23.0846659Z Test results will be stored in test-reports/python-pytest/inductor.test_compiled_autograd/inductor.test_compiled_autograd-857de25fa6c3d7b5.xml 2025-12-04T09:38:23.0847488Z ============================= test session starts ============================== 2025-12-04T09:38:23.0848089Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:38:23.0848595Z cachedir: .pytest_cache 2025-12-04T09:38:23.0849701Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:38:23.0850466Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:38:23.0850861Z configfile: pytest.ini 2025-12-04T09:38:23.0851515Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:38:23.0852247Z collecting ... collected 861 items 2025-12-04T09:38:23.0852794Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:38:23.0880948Z Running 50 items in this shard: test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.0906756Z 2025-12-04T09:38:23.0907744Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py W1204 09:38:14.721000 11686 site-packages/torch/_inductor/utils.py:2565] [!0/0/0] DeviceCopy in input program 2025-12-04T09:38:23.0909068Z W1204 09:38:16.711000 11686 site-packages/torch/_inductor/utils.py:2565] [!1/6/0] DeviceCopy in input program 2025-12-04T09:38:23.0909590Z PASSED [7.8595s] [ 2%] 2025-12-04T09:38:23.0910360Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0257s] [ 2%] 2025-12-04T09:38:23.0911691Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0011s] [ 2%] 2025-12-04T09:38:23.0913011Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0914422Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0011s] [ 2%] 2025-12-04T09:38:23.0915735Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0917050Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0918559Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0919868Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0921191Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0922501Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0923824Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0925138Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0926448Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0927959Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0929286Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0930609Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0010s] [ 2%] 2025-12-04T09:38:23.0931935Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0942111Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0943444Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0944763Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0946071Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0947391Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0016s] [ 2%] 2025-12-04T09:38:23.0948703Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0950013Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0951483Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0952799Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0954115Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0955419Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0010s] [ 2%] 2025-12-04T09:38:23.0956730Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0958103Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0959403Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0960708Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0962015Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0963321Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0010s] [ 2%] 2025-12-04T09:38:23.0965717Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0967032Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0009s] [ 2%] 2025-12-04T09:38:23.0968410Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0969727Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0971044Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0972369Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0010s] [ 2%] 2025-12-04T09:38:23.0973699Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0975016Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0976326Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0977641Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0978948Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0980343Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0010s] [ 2%] 2025-12-04T09:38:23.0981660Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0982974Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0984275Z inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel <- test/test_autograd.py FAILED [0.0008s] [ 2%] 2025-12-04T09:38:23.0985000Z 2025-12-04T09:38:23.0985130Z =================================== FAILURES =================================== 2025-12-04T09:38:23.0985680Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.0986218Z Traceback (most recent call last): 2025-12-04T09:38:23.0986834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.0987441Z method(*args, **kwargs) 2025-12-04T09:38:23.0987940Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.0988468Z stack.enter_context(ctx) 2025-12-04T09:38:23.0988901Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.0989374Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.0989816Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.0990352Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.0990650Z AttributeError: args 2025-12-04T09:38:23.0990803Z 2025-12-04T09:38:23.0990993Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.0991938Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.0992676Z 2025-12-04T09:38:23.0992899Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.0993417Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.0994875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.0996169Z warnings.warn( 2025-12-04T09:38:23.0998312Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1000479Z b_grad = a.grad 2025-12-04T09:38:23.1002581Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1004838Z c_grad = a.grad 2025-12-04T09:38:23.1005286Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1005809Z Traceback (most recent call last): 2025-12-04T09:38:23.1006412Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1007019Z method(*args, **kwargs) 2025-12-04T09:38:23.1007565Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1008086Z stack.enter_context(ctx) 2025-12-04T09:38:23.1008521Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1008997Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1009436Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1009887Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1010192Z AttributeError: args 2025-12-04T09:38:23.1010343Z 2025-12-04T09:38:23.1010533Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1011457Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1012205Z 2025-12-04T09:38:23.1012428Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1012939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1014194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1015440Z warnings.warn( 2025-12-04T09:38:23.1017554Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1019690Z b_grad = a.grad 2025-12-04T09:38:23.1021777Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1023938Z c_grad = a.grad 2025-12-04T09:38:23.1024383Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1024899Z Traceback (most recent call last): 2025-12-04T09:38:23.1025513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1026125Z method(*args, **kwargs) 2025-12-04T09:38:23.1026621Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1027158Z stack.enter_context(ctx) 2025-12-04T09:38:23.1027641Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1028113Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1028631Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1029095Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1029393Z AttributeError: args 2025-12-04T09:38:23.1029544Z 2025-12-04T09:38:23.1029731Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1030654Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1031405Z 2025-12-04T09:38:23.1031625Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1032131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1033400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1034547Z warnings.warn( 2025-12-04T09:38:23.1036812Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1039025Z b_grad = a.grad 2025-12-04T09:38:23.1041261Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1043431Z c_grad = a.grad 2025-12-04T09:38:23.1043881Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1044399Z Traceback (most recent call last): 2025-12-04T09:38:23.1045008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1045617Z method(*args, **kwargs) 2025-12-04T09:38:23.1046109Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1046644Z stack.enter_context(ctx) 2025-12-04T09:38:23.1047083Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1047596Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1048027Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1048488Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1048787Z AttributeError: args 2025-12-04T09:38:23.1048939Z 2025-12-04T09:38:23.1049117Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1050051Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1050802Z 2025-12-04T09:38:23.1051027Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1051540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1052923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1054080Z warnings.warn( 2025-12-04T09:38:23.1056188Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1058364Z b_grad = a.grad 2025-12-04T09:38:23.1060469Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1062624Z c_grad = a.grad 2025-12-04T09:38:23.1063062Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1063584Z Traceback (most recent call last): 2025-12-04T09:38:23.1064274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1064883Z method(*args, **kwargs) 2025-12-04T09:38:23.1065379Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1065909Z stack.enter_context(ctx) 2025-12-04T09:38:23.1066345Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1066807Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1067243Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1067702Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1068005Z AttributeError: args 2025-12-04T09:38:23.1068158Z 2025-12-04T09:38:23.1068338Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1069269Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1070020Z 2025-12-04T09:38:23.1070239Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1070753Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1072014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1073164Z warnings.warn( 2025-12-04T09:38:23.1075346Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1077485Z b_grad = a.grad 2025-12-04T09:38:23.1079581Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1081703Z c_grad = a.grad 2025-12-04T09:38:23.1082143Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1082671Z Traceback (most recent call last): 2025-12-04T09:38:23.1083282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1083890Z method(*args, **kwargs) 2025-12-04T09:38:23.1084376Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1084908Z stack.enter_context(ctx) 2025-12-04T09:38:23.1085345Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1085808Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1086256Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1086718Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1087010Z AttributeError: args 2025-12-04T09:38:23.1087169Z 2025-12-04T09:38:23.1087348Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1088418Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1089160Z 2025-12-04T09:38:23.1089394Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1089897Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1091157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1092318Z warnings.warn( 2025-12-04T09:38:23.1094435Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1096578Z b_grad = a.grad 2025-12-04T09:38:23.1098673Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1100841Z c_grad = a.grad 2025-12-04T09:38:23.1101280Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1101903Z Traceback (most recent call last): 2025-12-04T09:38:23.1102519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1103125Z method(*args, **kwargs) 2025-12-04T09:38:23.1103624Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1104160Z stack.enter_context(ctx) 2025-12-04T09:38:23.1104595Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1105052Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1105493Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1105959Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1106249Z AttributeError: args 2025-12-04T09:38:23.1106410Z 2025-12-04T09:38:23.1106593Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1107533Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1108279Z 2025-12-04T09:38:23.1108511Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1109017Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1110272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1111425Z warnings.warn( 2025-12-04T09:38:23.1113643Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1115772Z b_grad = a.grad 2025-12-04T09:38:23.1117873Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1120039Z c_grad = a.grad 2025-12-04T09:38:23.1120482Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1121001Z Traceback (most recent call last): 2025-12-04T09:38:23.1121604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1122217Z method(*args, **kwargs) 2025-12-04T09:38:23.1122725Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1123263Z stack.enter_context(ctx) 2025-12-04T09:38:23.1123693Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1124168Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1124614Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1125072Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1125448Z AttributeError: args 2025-12-04T09:38:23.1125617Z 2025-12-04T09:38:23.1125796Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1126732Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1127506Z 2025-12-04T09:38:23.1127760Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1128289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1129540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1130702Z warnings.warn( 2025-12-04T09:38:23.1132821Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1134952Z b_grad = a.grad 2025-12-04T09:38:23.1137165Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1139436Z c_grad = a.grad 2025-12-04T09:38:23.1139882Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1140407Z Traceback (most recent call last): 2025-12-04T09:38:23.1141004Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1141614Z method(*args, **kwargs) 2025-12-04T09:38:23.1142115Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1142643Z stack.enter_context(ctx) 2025-12-04T09:38:23.1143081Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1143553Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1143998Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1144447Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1144739Z AttributeError: args 2025-12-04T09:38:23.1144889Z 2025-12-04T09:38:23.1145077Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1146002Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1146738Z 2025-12-04T09:38:23.1146960Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1147468Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1148865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1150016Z warnings.warn( 2025-12-04T09:38:23.1152126Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1154254Z b_grad = a.grad 2025-12-04T09:38:23.1156355Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1158524Z c_grad = a.grad 2025-12-04T09:38:23.1158975Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1159497Z Traceback (most recent call last): 2025-12-04T09:38:23.1160103Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1160708Z method(*args, **kwargs) 2025-12-04T09:38:23.1161292Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1161821Z stack.enter_context(ctx) 2025-12-04T09:38:23.1162263Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1162733Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1163171Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1163624Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1163923Z AttributeError: args 2025-12-04T09:38:23.1164075Z 2025-12-04T09:38:23.1164269Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1165191Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1165941Z 2025-12-04T09:38:23.1166163Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1166680Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1168025Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1169194Z warnings.warn( 2025-12-04T09:38:23.1171310Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1173445Z b_grad = a.grad 2025-12-04T09:38:23.1175622Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1177758Z c_grad = a.grad 2025-12-04T09:38:23.1178205Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1178731Z Traceback (most recent call last): 2025-12-04T09:38:23.1179340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1179964Z method(*args, **kwargs) 2025-12-04T09:38:23.1180474Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1181010Z stack.enter_context(ctx) 2025-12-04T09:38:23.1181458Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1181940Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1182381Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1182859Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1183169Z AttributeError: args 2025-12-04T09:38:23.1183326Z 2025-12-04T09:38:23.1183521Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1184454Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1185291Z 2025-12-04T09:38:23.1185514Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1186031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1187297Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1188461Z warnings.warn( 2025-12-04T09:38:23.1190583Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1192734Z b_grad = a.grad 2025-12-04T09:38:23.1194825Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1196959Z c_grad = a.grad 2025-12-04T09:38:23.1197407Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1197949Z Traceback (most recent call last): 2025-12-04T09:38:23.1198663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1199266Z method(*args, **kwargs) 2025-12-04T09:38:23.1199764Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1200287Z stack.enter_context(ctx) 2025-12-04T09:38:23.1200716Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1201178Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1201603Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1202056Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1202352Z AttributeError: args 2025-12-04T09:38:23.1202503Z 2025-12-04T09:38:23.1202681Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1203615Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1204367Z 2025-12-04T09:38:23.1204590Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1205092Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1206347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1207587Z warnings.warn( 2025-12-04T09:38:23.1209699Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1211939Z b_grad = a.grad 2025-12-04T09:38:23.1214038Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1216170Z c_grad = a.grad 2025-12-04T09:38:23.1216624Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1217149Z Traceback (most recent call last): 2025-12-04T09:38:23.1217762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1218368Z method(*args, **kwargs) 2025-12-04T09:38:23.1218861Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1219394Z stack.enter_context(ctx) 2025-12-04T09:38:23.1219836Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1220304Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1220743Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1221211Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1221519Z AttributeError: args 2025-12-04T09:38:23.1221672Z 2025-12-04T09:38:23.1221853Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1222867Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1223613Z 2025-12-04T09:38:23.1223831Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1224337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1225586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1226740Z warnings.warn( 2025-12-04T09:38:23.1228865Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1231026Z b_grad = a.grad 2025-12-04T09:38:23.1233121Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1235487Z c_grad = a.grad 2025-12-04T09:38:23.1235921Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1236448Z Traceback (most recent call last): 2025-12-04T09:38:23.1237061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1237672Z method(*args, **kwargs) 2025-12-04T09:38:23.1238221Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1238759Z stack.enter_context(ctx) 2025-12-04T09:38:23.1239196Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1239666Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1240114Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1240575Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1240886Z AttributeError: args 2025-12-04T09:38:23.1241047Z 2025-12-04T09:38:23.1241226Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1242166Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1242907Z 2025-12-04T09:38:23.1243138Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1243658Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1244907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1246062Z warnings.warn( 2025-12-04T09:38:23.1248388Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1250540Z b_grad = a.grad 2025-12-04T09:38:23.1252646Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1254793Z c_grad = a.grad 2025-12-04T09:38:23.1255244Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1255779Z Traceback (most recent call last): 2025-12-04T09:38:23.1256400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1257008Z method(*args, **kwargs) 2025-12-04T09:38:23.1257501Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1258037Z stack.enter_context(ctx) 2025-12-04T09:38:23.1258585Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1259042Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1259491Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1259951Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1260237Z AttributeError: args 2025-12-04T09:38:23.1260390Z 2025-12-04T09:38:23.1260565Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1261487Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1262225Z 2025-12-04T09:38:23.1262448Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1262948Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1264218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1265366Z warnings.warn( 2025-12-04T09:38:23.1267473Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1269681Z b_grad = a.grad 2025-12-04T09:38:23.1271869Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1274014Z c_grad = a.grad 2025-12-04T09:38:23.1274455Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1274978Z Traceback (most recent call last): 2025-12-04T09:38:23.1275593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1276203Z method(*args, **kwargs) 2025-12-04T09:38:23.1276713Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1277253Z stack.enter_context(ctx) 2025-12-04T09:38:23.1277701Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1278201Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1278634Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1279095Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1279391Z AttributeError: args 2025-12-04T09:38:23.1279548Z 2025-12-04T09:38:23.1279725Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1280652Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1281385Z 2025-12-04T09:38:23.1281607Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1282190Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1283438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1284585Z warnings.warn( 2025-12-04T09:38:23.1286694Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1288877Z b_grad = a.grad 2025-12-04T09:38:23.1290956Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1293109Z c_grad = a.grad 2025-12-04T09:38:23.1299070Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1299628Z Traceback (most recent call last): 2025-12-04T09:38:23.1300251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1300870Z method(*args, **kwargs) 2025-12-04T09:38:23.1301490Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1302029Z stack.enter_context(ctx) 2025-12-04T09:38:23.1302456Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1302933Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1303368Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1303826Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1304134Z AttributeError: args 2025-12-04T09:38:23.1304297Z 2025-12-04T09:38:23.1304480Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1305407Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1306150Z 2025-12-04T09:38:23.1306371Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1306881Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1308196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1309354Z warnings.warn( 2025-12-04T09:38:23.1311466Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1313691Z b_grad = a.grad 2025-12-04T09:38:23.1315798Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1317974Z c_grad = a.grad 2025-12-04T09:38:23.1318421Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1318955Z Traceback (most recent call last): 2025-12-04T09:38:23.1319565Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1320182Z method(*args, **kwargs) 2025-12-04T09:38:23.1320682Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1321220Z stack.enter_context(ctx) 2025-12-04T09:38:23.1321662Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1322136Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1322582Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1323030Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1323332Z AttributeError: args 2025-12-04T09:38:23.1323488Z 2025-12-04T09:38:23.1323678Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1324608Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1325442Z 2025-12-04T09:38:23.1325666Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1326178Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1327527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1328705Z warnings.warn( 2025-12-04T09:38:23.1330814Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1332984Z b_grad = a.grad 2025-12-04T09:38:23.1335070Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1337521Z c_grad = a.grad 2025-12-04T09:38:23.1337967Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1338496Z Traceback (most recent call last): 2025-12-04T09:38:23.1339097Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1339698Z method(*args, **kwargs) 2025-12-04T09:38:23.1340187Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1340708Z stack.enter_context(ctx) 2025-12-04T09:38:23.1341143Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1341611Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1342040Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1342499Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1342801Z AttributeError: args 2025-12-04T09:38:23.1342953Z 2025-12-04T09:38:23.1343140Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1344068Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1344815Z 2025-12-04T09:38:23.1345034Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1345531Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1346786Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1347978Z warnings.warn( 2025-12-04T09:38:23.1350221Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1352353Z b_grad = a.grad 2025-12-04T09:38:23.1354441Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1356571Z c_grad = a.grad 2025-12-04T09:38:23.1357012Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1357528Z Traceback (most recent call last): 2025-12-04T09:38:23.1358135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1358738Z method(*args, **kwargs) 2025-12-04T09:38:23.1359233Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1359756Z stack.enter_context(ctx) 2025-12-04T09:38:23.1360190Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1360656Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1361194Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1361646Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1361941Z AttributeError: args 2025-12-04T09:38:23.1362098Z 2025-12-04T09:38:23.1362274Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1363196Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1363943Z 2025-12-04T09:38:23.1364160Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1364659Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1365903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1367040Z warnings.warn( 2025-12-04T09:38:23.1369239Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1371367Z b_grad = a.grad 2025-12-04T09:38:23.1373531Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1375658Z c_grad = a.grad 2025-12-04T09:38:23.1376094Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1376601Z Traceback (most recent call last): 2025-12-04T09:38:23.1377202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1377846Z method(*args, **kwargs) 2025-12-04T09:38:23.1378337Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1378857Z stack.enter_context(ctx) 2025-12-04T09:38:23.1379285Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1379740Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1380170Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1380622Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1380916Z AttributeError: args 2025-12-04T09:38:23.1381067Z 2025-12-04T09:38:23.1381245Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1382169Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1382915Z 2025-12-04T09:38:23.1383135Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1383637Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1384981Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1386133Z warnings.warn( 2025-12-04T09:38:23.1388241Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1390360Z b_grad = a.grad 2025-12-04T09:38:23.1392450Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1394577Z c_grad = a.grad 2025-12-04T09:38:23.1395009Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1395522Z Traceback (most recent call last): 2025-12-04T09:38:23.1396129Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1396731Z method(*args, **kwargs) 2025-12-04T09:38:23.1397227Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1397753Z stack.enter_context(ctx) 2025-12-04T09:38:23.1398316Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1398769Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1399195Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1399648Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1399943Z AttributeError: args 2025-12-04T09:38:23.1400094Z 2025-12-04T09:38:23.1400272Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1401324Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1402058Z 2025-12-04T09:38:23.1402279Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1402784Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1404030Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1405175Z warnings.warn( 2025-12-04T09:38:23.1407278Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1409635Z b_grad = a.grad 2025-12-04T09:38:23.1411722Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1413882Z c_grad = a.grad 2025-12-04T09:38:23.1414315Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1414843Z Traceback (most recent call last): 2025-12-04T09:38:23.1415449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1416051Z method(*args, **kwargs) 2025-12-04T09:38:23.1416537Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1417061Z stack.enter_context(ctx) 2025-12-04T09:38:23.1417485Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1417937Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1418359Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1418809Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1419091Z AttributeError: args 2025-12-04T09:38:23.1419246Z 2025-12-04T09:38:23.1419423Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1420342Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1421087Z 2025-12-04T09:38:23.1421310Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1421910Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1423161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1424305Z warnings.warn( 2025-12-04T09:38:23.1426410Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1428584Z b_grad = a.grad 2025-12-04T09:38:23.1430672Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1432784Z c_grad = a.grad 2025-12-04T09:38:23.1433209Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1433804Z Traceback (most recent call last): 2025-12-04T09:38:23.1434400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1435012Z method(*args, **kwargs) 2025-12-04T09:38:23.1435635Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1436156Z stack.enter_context(ctx) 2025-12-04T09:38:23.1436583Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1437043Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1437519Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1437962Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1438248Z AttributeError: args 2025-12-04T09:38:23.1438402Z 2025-12-04T09:38:23.1438577Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1439512Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1440265Z 2025-12-04T09:38:23.1440487Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1440976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1442220Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1443361Z warnings.warn( 2025-12-04T09:38:23.1445576Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1447794Z b_grad = a.grad 2025-12-04T09:38:23.1449875Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1452025Z c_grad = a.grad 2025-12-04T09:38:23.1452460Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1452986Z Traceback (most recent call last): 2025-12-04T09:38:23.1453593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1454194Z method(*args, **kwargs) 2025-12-04T09:38:23.1454687Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1455209Z stack.enter_context(ctx) 2025-12-04T09:38:23.1455628Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1456089Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1456518Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1456962Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1457360Z AttributeError: args 2025-12-04T09:38:23.1457507Z 2025-12-04T09:38:23.1457689Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1458612Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1459348Z 2025-12-04T09:38:23.1459563Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1460056Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1461311Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1462455Z warnings.warn( 2025-12-04T09:38:23.1464556Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1466704Z b_grad = a.grad 2025-12-04T09:38:23.1468839Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1470961Z c_grad = a.grad 2025-12-04T09:38:23.1471472Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1471992Z Traceback (most recent call last): 2025-12-04T09:38:23.1472589Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1473182Z method(*args, **kwargs) 2025-12-04T09:38:23.1473672Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1474191Z stack.enter_context(ctx) 2025-12-04T09:38:23.1474608Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1475065Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1475497Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1475941Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1476225Z AttributeError: args 2025-12-04T09:38:23.1476379Z 2025-12-04T09:38:23.1476560Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1477484Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1478277Z 2025-12-04T09:38:23.1478497Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1478993Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1480238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1481456Z warnings.warn( 2025-12-04T09:38:23.1483554Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1485669Z b_grad = a.grad 2025-12-04T09:38:23.1487797Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1489925Z c_grad = a.grad 2025-12-04T09:38:23.1490353Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1490869Z Traceback (most recent call last): 2025-12-04T09:38:23.1491459Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1492059Z method(*args, **kwargs) 2025-12-04T09:38:23.1492546Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1493063Z stack.enter_context(ctx) 2025-12-04T09:38:23.1493492Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1493949Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1494458Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1494908Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1495195Z AttributeError: args 2025-12-04T09:38:23.1495343Z 2025-12-04T09:38:23.1495520Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1496429Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1497172Z 2025-12-04T09:38:23.1497390Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1497890Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1499203Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1500339Z warnings.warn( 2025-12-04T09:38:23.1502428Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1504579Z b_grad = a.grad 2025-12-04T09:38:23.1506661Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1508906Z c_grad = a.grad 2025-12-04T09:38:23.1509339Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1509846Z Traceback (most recent call last): 2025-12-04T09:38:23.1510442Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1511040Z method(*args, **kwargs) 2025-12-04T09:38:23.1511532Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1512050Z stack.enter_context(ctx) 2025-12-04T09:38:23.1512484Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1512948Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1513368Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1513817Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1514105Z AttributeError: args 2025-12-04T09:38:23.1514254Z 2025-12-04T09:38:23.1514431Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1515344Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1516090Z 2025-12-04T09:38:23.1516307Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1516804Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1518161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1519305Z warnings.warn( 2025-12-04T09:38:23.1521407Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1523534Z b_grad = a.grad 2025-12-04T09:38:23.1525615Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1527800Z c_grad = a.grad 2025-12-04T09:38:23.1528279Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1528793Z Traceback (most recent call last): 2025-12-04T09:38:23.1529393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1530076Z method(*args, **kwargs) 2025-12-04T09:38:23.1530574Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1531089Z stack.enter_context(ctx) 2025-12-04T09:38:23.1531515Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1531973Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1532395Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1532845Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1533135Z AttributeError: args 2025-12-04T09:38:23.1533282Z 2025-12-04T09:38:23.1533459Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1534374Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1535120Z 2025-12-04T09:38:23.1535502Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1536007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1537251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1538442Z warnings.warn( 2025-12-04T09:38:23.1540540Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1542785Z b_grad = a.grad 2025-12-04T09:38:23.1544873Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1546992Z c_grad = a.grad 2025-12-04T09:38:23.1547425Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1547939Z Traceback (most recent call last): 2025-12-04T09:38:23.1548537Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1549140Z method(*args, **kwargs) 2025-12-04T09:38:23.1549621Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1550145Z stack.enter_context(ctx) 2025-12-04T09:38:23.1550575Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1551030Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1551449Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1551895Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1552183Z AttributeError: args 2025-12-04T09:38:23.1552333Z 2025-12-04T09:38:23.1552513Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1553548Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1554287Z 2025-12-04T09:38:23.1554512Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1555008Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1556244Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1557377Z warnings.warn( 2025-12-04T09:38:23.1559527Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1561677Z b_grad = a.grad 2025-12-04T09:38:23.1563759Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1565878Z c_grad = a.grad 2025-12-04T09:38:23.1566307Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1566894Z Traceback (most recent call last): 2025-12-04T09:38:23.1567584Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1568180Z method(*args, **kwargs) 2025-12-04T09:38:23.1568660Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1569178Z stack.enter_context(ctx) 2025-12-04T09:38:23.1569602Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1570055Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1570485Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1570932Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1571220Z AttributeError: args 2025-12-04T09:38:23.1571370Z 2025-12-04T09:38:23.1571548Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1572471Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1573214Z 2025-12-04T09:38:23.1573434Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1573926Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1575168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1576310Z warnings.warn( 2025-12-04T09:38:23.1578406Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1580610Z b_grad = a.grad 2025-12-04T09:38:23.1582695Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1584812Z c_grad = a.grad 2025-12-04T09:38:23.1585242Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1585757Z Traceback (most recent call last): 2025-12-04T09:38:23.1586350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1586940Z method(*args, **kwargs) 2025-12-04T09:38:23.1587424Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1587944Z stack.enter_context(ctx) 2025-12-04T09:38:23.1588421Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1588875Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1589307Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1589754Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1590042Z AttributeError: args 2025-12-04T09:38:23.1590198Z 2025-12-04T09:38:23.1590456Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1591370Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1592114Z 2025-12-04T09:38:23.1592333Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1592823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1594063Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1595205Z warnings.warn( 2025-12-04T09:38:23.1597309Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1599478Z b_grad = a.grad 2025-12-04T09:38:23.1601558Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1603803Z c_grad = a.grad 2025-12-04T09:38:23.1604236Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1604749Z Traceback (most recent call last): 2025-12-04T09:38:23.1605346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1605936Z method(*args, **kwargs) 2025-12-04T09:38:23.1606421Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1606948Z stack.enter_context(ctx) 2025-12-04T09:38:23.1607428Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1607886Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1608312Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1608762Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1609043Z AttributeError: args 2025-12-04T09:38:23.1609195Z 2025-12-04T09:38:23.1609373Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1610284Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1611018Z 2025-12-04T09:38:23.1611237Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1611727Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1612969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1614193Z warnings.warn( 2025-12-04T09:38:23.1616296Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1618502Z b_grad = a.grad 2025-12-04T09:38:23.1620587Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1622730Z c_grad = a.grad 2025-12-04T09:38:23.1623160Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1623669Z Traceback (most recent call last): 2025-12-04T09:38:23.1624265Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1624859Z method(*args, **kwargs) 2025-12-04T09:38:23.1625342Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1625954Z stack.enter_context(ctx) 2025-12-04T09:38:23.1626371Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1626830Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1627249Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1627699Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1628019Z AttributeError: args 2025-12-04T09:38:23.1628166Z 2025-12-04T09:38:23.1628343Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1629266Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1634195Z 2025-12-04T09:38:23.1634440Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1634955Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1636376Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1637515Z warnings.warn( 2025-12-04T09:38:23.1639598Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1641712Z b_grad = a.grad 2025-12-04T09:38:23.1643936Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1646060Z c_grad = a.grad 2025-12-04T09:38:23.1646501Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1647013Z Traceback (most recent call last): 2025-12-04T09:38:23.1647658Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1648258Z method(*args, **kwargs) 2025-12-04T09:38:23.1648744Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1649268Z stack.enter_context(ctx) 2025-12-04T09:38:23.1649689Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1650145Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1650569Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1651014Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1651304Z AttributeError: args 2025-12-04T09:38:23.1651452Z 2025-12-04T09:38:23.1651634Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1652548Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1653399Z 2025-12-04T09:38:23.1653616Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1654115Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1655354Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1656492Z warnings.warn( 2025-12-04T09:38:23.1658577Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1660723Z b_grad = a.grad 2025-12-04T09:38:23.1662796Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1664893Z c_grad = a.grad 2025-12-04T09:38:23.1665321Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1665839Z Traceback (most recent call last): 2025-12-04T09:38:23.1666433Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1667107Z method(*args, **kwargs) 2025-12-04T09:38:23.1667591Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1668105Z stack.enter_context(ctx) 2025-12-04T09:38:23.1668524Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1668978Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1669402Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1669845Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1670136Z AttributeError: args 2025-12-04T09:38:23.1670283Z 2025-12-04T09:38:23.1670462Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1671378Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1672115Z 2025-12-04T09:38:23.1672342Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1672837Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1674070Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1675197Z warnings.warn( 2025-12-04T09:38:23.1677328Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1679529Z b_grad = a.grad 2025-12-04T09:38:23.1681593Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1683744Z c_grad = a.grad 2025-12-04T09:38:23.1684173Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1684681Z Traceback (most recent call last): 2025-12-04T09:38:23.1685278Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1685872Z method(*args, **kwargs) 2025-12-04T09:38:23.1686353Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1686864Z stack.enter_context(ctx) 2025-12-04T09:38:23.1687311Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1687838Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1688255Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1688698Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1688988Z AttributeError: args 2025-12-04T09:38:23.1689134Z 2025-12-04T09:38:23.1689309Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1690306Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1691041Z 2025-12-04T09:38:23.1691256Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1691748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1692984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1694120Z warnings.warn( 2025-12-04T09:38:23.1696215Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1698322Z b_grad = a.grad 2025-12-04T09:38:23.1700386Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1702622Z c_grad = a.grad 2025-12-04T09:38:23.1703059Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1703566Z Traceback (most recent call last): 2025-12-04T09:38:23.1704159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1704748Z method(*args, **kwargs) 2025-12-04T09:38:23.1705223Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1705740Z stack.enter_context(ctx) 2025-12-04T09:38:23.1706159Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1706613Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1707049Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1707537Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1707824Z AttributeError: args 2025-12-04T09:38:23.1707971Z 2025-12-04T09:38:23.1708147Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1709059Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1709797Z 2025-12-04T09:38:23.1710013Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1710502Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1711736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1712874Z warnings.warn( 2025-12-04T09:38:23.1715041Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1717172Z b_grad = a.grad 2025-12-04T09:38:23.1719284Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1721419Z c_grad = a.grad 2025-12-04T09:38:23.1721845Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1722352Z Traceback (most recent call last): 2025-12-04T09:38:23.1722944Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1723536Z method(*args, **kwargs) 2025-12-04T09:38:23.1724013Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1724529Z stack.enter_context(ctx) 2025-12-04T09:38:23.1725035Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1725487Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1725913Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1726355Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1726638Z AttributeError: args 2025-12-04T09:38:23.1726787Z 2025-12-04T09:38:23.1726963Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1727924Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1728660Z 2025-12-04T09:38:23.1728877Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1729364Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1730602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1731760Z warnings.warn( 2025-12-04T09:38:23.1733847Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1736098Z b_grad = a.grad 2025-12-04T09:38:23.1738291Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1740417Z c_grad = a.grad 2025-12-04T09:38:23.1740844Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1741356Z Traceback (most recent call last): 2025-12-04T09:38:23.1741954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1742544Z method(*args, **kwargs) 2025-12-04T09:38:23.1743027Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1743549Z stack.enter_context(ctx) 2025-12-04T09:38:23.1743970Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1744424Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1744848Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1745291Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1745573Z AttributeError: args 2025-12-04T09:38:23.1745726Z 2025-12-04T09:38:23.1745901Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1746819Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1747556Z 2025-12-04T09:38:23.1747775Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1748378Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1749627Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1750761Z warnings.warn( 2025-12-04T09:38:23.1752851Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1754999Z b_grad = a.grad 2025-12-04T09:38:23.1757085Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1759222Z c_grad = a.grad 2025-12-04T09:38:23.1759650Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1760157Z Traceback (most recent call last): 2025-12-04T09:38:23.1760748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1761345Z method(*args, **kwargs) 2025-12-04T09:38:23.1761825Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1762432Z stack.enter_context(ctx) 2025-12-04T09:38:23.1762851Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1763301Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1763721Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1764169Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1764450Z AttributeError: args 2025-12-04T09:38:23.1764602Z 2025-12-04T09:38:23.1764777Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1765695Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1766432Z 2025-12-04T09:38:23.1766649Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1767140Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1768640Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1770010Z warnings.warn( 2025-12-04T09:38:23.1772554Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1775217Z b_grad = a.grad 2025-12-04T09:38:23.1777749Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1780326Z c_grad = a.grad 2025-12-04T09:38:23.1780809Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1781390Z Traceback (most recent call last): 2025-12-04T09:38:23.1782067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1782758Z method(*args, **kwargs) 2025-12-04T09:38:23.1783300Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1783889Z stack.enter_context(ctx) 2025-12-04T09:38:23.1784354Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1784870Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1785343Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1785844Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1786151Z AttributeError: args 2025-12-04T09:38:23.1786310Z 2025-12-04T09:38:23.1786498Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1787629Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1788515Z 2025-12-04T09:38:23.1788865Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1789430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1790923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1792292Z warnings.warn( 2025-12-04T09:38:23.1794842Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1797420Z b_grad = a.grad 2025-12-04T09:38:23.1799999Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1802652Z c_grad = a.grad 2025-12-04T09:38:23.1803134Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1803713Z Traceback (most recent call last): 2025-12-04T09:38:23.1804400Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1805086Z method(*args, **kwargs) 2025-12-04T09:38:23.1805627Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1806214Z stack.enter_context(ctx) 2025-12-04T09:38:23.1806680Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1807196Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1807680Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1808120Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1808409Z AttributeError: args 2025-12-04T09:38:23.1808557Z 2025-12-04T09:38:23.1808735Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1809650Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1810390Z 2025-12-04T09:38:23.1810605Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1811099Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1812342Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1813475Z warnings.warn( 2025-12-04T09:38:23.1815649Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1817801Z b_grad = a.grad 2025-12-04T09:38:23.1819885Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1822032Z c_grad = a.grad 2025-12-04T09:38:23.1822468Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1822978Z Traceback (most recent call last): 2025-12-04T09:38:23.1823569Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1824163Z method(*args, **kwargs) 2025-12-04T09:38:23.1824642Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1825157Z stack.enter_context(ctx) 2025-12-04T09:38:23.1825578Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1826034Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1826536Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1826981Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1827264Z AttributeError: args 2025-12-04T09:38:23.1827410Z 2025-12-04T09:38:23.1827608Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1828554Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1829290Z 2025-12-04T09:38:23.1829505Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1829998Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1831241Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1832378Z warnings.warn( 2025-12-04T09:38:23.1834483Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1836768Z b_grad = a.grad 2025-12-04T09:38:23.1839025Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1841173Z c_grad = a.grad 2025-12-04T09:38:23.1841610Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1842113Z Traceback (most recent call last): 2025-12-04T09:38:23.1842704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1843296Z method(*args, **kwargs) 2025-12-04T09:38:23.1843774Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1844286Z stack.enter_context(ctx) 2025-12-04T09:38:23.1844707Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1845169Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1845589Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1846039Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1846322Z AttributeError: args 2025-12-04T09:38:23.1846469Z 2025-12-04T09:38:23.1846649Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1847604Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1848345Z 2025-12-04T09:38:23.1848560Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1849052Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1850291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1851556Z warnings.warn( 2025-12-04T09:38:23.1853653Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1855772Z b_grad = a.grad 2025-12-04T09:38:23.1857859Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1860012Z c_grad = a.grad 2025-12-04T09:38:23.1860441Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1860950Z Traceback (most recent call last): 2025-12-04T09:38:23.1861545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1862141Z method(*args, **kwargs) 2025-12-04T09:38:23.1862619Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1863144Z stack.enter_context(ctx) 2025-12-04T09:38:23.1863563Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1864098Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1864522Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1864968Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1865255Z AttributeError: args 2025-12-04T09:38:23.1865401Z 2025-12-04T09:38:23.1865577Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1866492Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1867232Z 2025-12-04T09:38:23.1867448Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1867953Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1868922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1869008Z warnings.warn( 2025-12-04T09:38:23.1870967Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1871124Z b_grad = a.grad 2025-12-04T09:38:23.1873081Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1873160Z c_grad = a.grad 2025-12-04T09:38:23.1873463Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1873566Z Traceback (most recent call last): 2025-12-04T09:38:23.1873976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1874070Z method(*args, **kwargs) 2025-12-04T09:38:23.1874389Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1874493Z stack.enter_context(ctx) 2025-12-04T09:38:23.1874743Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1874839Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1875079Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1875182Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1875273Z AttributeError: args 2025-12-04T09:38:23.1875278Z 2025-12-04T09:38:23.1875456Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1876099Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1876112Z 2025-12-04T09:38:23.1876331Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1876612Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1877585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1877667Z warnings.warn( 2025-12-04T09:38:23.1879630Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1879713Z b_grad = a.grad 2025-12-04T09:38:23.1881661Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1881743Z c_grad = a.grad 2025-12-04T09:38:23.1882045Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1882229Z Traceback (most recent call last): 2025-12-04T09:38:23.1882641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1882735Z method(*args, **kwargs) 2025-12-04T09:38:23.1883056Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1883149Z stack.enter_context(ctx) 2025-12-04T09:38:23.1883401Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1883500Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1883734Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1883841Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1883927Z AttributeError: args 2025-12-04T09:38:23.1883932Z 2025-12-04T09:38:23.1884107Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1884750Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1884754Z 2025-12-04T09:38:23.1884978Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1885154Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1886112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1886196Z warnings.warn( 2025-12-04T09:38:23.1888288Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1888381Z b_grad = a.grad 2025-12-04T09:38:23.1890336Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1890418Z c_grad = a.grad 2025-12-04T09:38:23.1890722Z _ TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel _ 2025-12-04T09:38:23.1890823Z Traceback (most recent call last): 2025-12-04T09:38:23.1891240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3329, in wrapper 2025-12-04T09:38:23.1891331Z method(*args, **kwargs) 2025-12-04T09:38:23.1891651Z File "/var/lib/jenkins/workspace/test/inductor/test_compiled_autograd.py", line 5058, in wrapped 2025-12-04T09:38:23.1891749Z stack.enter_context(ctx) 2025-12-04T09:38:23.1892004Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 492, in enter_context 2025-12-04T09:38:23.1892101Z result = _cm_type.__enter__(cm) 2025-12-04T09:38:23.1892340Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 133, in __enter__ 2025-12-04T09:38:23.1892448Z del self.args, self.kwds, self.func 2025-12-04T09:38:23.1892616Z AttributeError: args 2025-12-04T09:38:23.1892620Z 2025-12-04T09:38:23.1892797Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1893437Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1893442Z 2025-12-04T09:38:23.1893661Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1893835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:38:23.1894794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:321: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance. 2025-12-04T09:38:23.1894876Z warnings.warn( 2025-12-04T09:38:23.1896839Z /var/lib/jenkins/workspace/test/test_autograd.py:7724: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1896925Z b_grad = a.grad 2025-12-04T09:38:23.1898873Z /var/lib/jenkins/workspace/test/test_autograd.py:7731: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T09:38:23.1898958Z c_grad = a.grad 2025-12-04T09:38:23.1899663Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_compiled_autograd/inductor.test_compiled_autograd-857de25fa6c3d7b5.xml - 2025-12-04T09:38:23.1899811Z =========================== short test summary info ============================ 2025-12-04T09:38:23.1900392Z FAILED [0.0257s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1900398Z 2025-12-04T09:38:23.1900577Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1901219Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1901228Z 2025-12-04T09:38:23.1901444Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1902025Z FAILED [0.0011s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1902029Z 2025-12-04T09:38:23.1902205Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1902836Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1902841Z 2025-12-04T09:38:23.1903055Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1903624Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1903704Z 2025-12-04T09:38:23.1903879Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1904513Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1904518Z 2025-12-04T09:38:23.1904735Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1905304Z FAILED [0.0011s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1905309Z 2025-12-04T09:38:23.1905486Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1906120Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1906130Z 2025-12-04T09:38:23.1906346Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1906915Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1906920Z 2025-12-04T09:38:23.1907094Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1907782Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1907786Z 2025-12-04T09:38:23.1908000Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1908569Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1908573Z 2025-12-04T09:38:23.1908749Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1909466Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1909474Z 2025-12-04T09:38:23.1909692Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1910255Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1910260Z 2025-12-04T09:38:23.1910437Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1911064Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1911069Z 2025-12-04T09:38:23.1911286Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1911857Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1911861Z 2025-12-04T09:38:23.1912039Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1912670Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1912674Z 2025-12-04T09:38:23.1912887Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1913456Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1913460Z 2025-12-04T09:38:23.1913635Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1914382Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1914386Z 2025-12-04T09:38:23.1914604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1915172Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1915177Z 2025-12-04T09:38:23.1915354Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1915987Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1915991Z 2025-12-04T09:38:23.1916208Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1916779Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1916783Z 2025-12-04T09:38:23.1916962Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1917642Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1917646Z 2025-12-04T09:38:23.1917858Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1918431Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1918435Z 2025-12-04T09:38:23.1918609Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1919236Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1919246Z 2025-12-04T09:38:23.1919536Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1920104Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1920111Z 2025-12-04T09:38:23.1920286Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1920920Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1920924Z 2025-12-04T09:38:23.1921143Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1921705Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1921714Z 2025-12-04T09:38:23.1921891Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1922523Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1922527Z 2025-12-04T09:38:23.1922741Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1923313Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1923317Z 2025-12-04T09:38:23.1923492Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1924128Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1924233Z 2025-12-04T09:38:23.1924450Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1925023Z FAILED [0.0010s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1925027Z 2025-12-04T09:38:23.1925201Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1925829Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1925833Z 2025-12-04T09:38:23.1926048Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1926617Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1926627Z 2025-12-04T09:38:23.1926805Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1927491Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1927495Z 2025-12-04T09:38:23.1927709Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1928276Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1928281Z 2025-12-04T09:38:23.1928456Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1929088Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1929098Z 2025-12-04T09:38:23.1929311Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1929961Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1929966Z 2025-12-04T09:38:23.1930142Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1930771Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1930776Z 2025-12-04T09:38:23.1930992Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1931560Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1931572Z 2025-12-04T09:38:23.1931751Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1932383Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1932387Z 2025-12-04T09:38:23.1932604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1933169Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1933174Z 2025-12-04T09:38:23.1933348Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1933984Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1933987Z 2025-12-04T09:38:23.1934288Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1934861Z FAILED [0.0016s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1934866Z 2025-12-04T09:38:23.1935040Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1935816Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1935824Z 2025-12-04T09:38:23.1936038Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1936605Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1936609Z 2025-12-04T09:38:23.1936791Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1937428Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1937436Z 2025-12-04T09:38:23.1937654Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1938216Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1938221Z 2025-12-04T09:38:23.1938396Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1939024Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1939028Z 2025-12-04T09:38:23.1939241Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1939816Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1939820Z 2025-12-04T09:38:23.1940112Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1940756Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1940760Z 2025-12-04T09:38:23.1940973Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1941544Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1941548Z 2025-12-04T09:38:23.1941731Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1942363Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1942374Z 2025-12-04T09:38:23.1942596Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1943161Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1943165Z 2025-12-04T09:38:23.1943343Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1943970Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1943974Z 2025-12-04T09:38:23.1944188Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1944765Z FAILED [0.0010s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1944872Z 2025-12-04T09:38:23.1945048Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1945696Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1945700Z 2025-12-04T09:38:23.1945913Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1946479Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1946486Z 2025-12-04T09:38:23.1946661Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1947285Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1947296Z 2025-12-04T09:38:23.1947512Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1948085Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1948090Z 2025-12-04T09:38:23.1948268Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1948902Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1948907Z 2025-12-04T09:38:23.1949121Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1949687Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1949696Z 2025-12-04T09:38:23.1949870Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1950583Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1950587Z 2025-12-04T09:38:23.1950804Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1951375Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1951379Z 2025-12-04T09:38:23.1951552Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1952178Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1952188Z 2025-12-04T09:38:23.1952404Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1952975Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1952979Z 2025-12-04T09:38:23.1953155Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1953782Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1953787Z 2025-12-04T09:38:23.1954000Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1954569Z FAILED [0.0010s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1954681Z 2025-12-04T09:38:23.1954856Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1955494Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1955498Z 2025-12-04T09:38:23.1955716Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1956301Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1956305Z 2025-12-04T09:38:23.1956480Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1957112Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1957122Z 2025-12-04T09:38:23.1957339Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1957907Z FAILED [0.0009s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1957911Z 2025-12-04T09:38:23.1958089Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1958717Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1958721Z 2025-12-04T09:38:23.1958937Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1959503Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1959507Z 2025-12-04T09:38:23.1959687Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1960404Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1960409Z 2025-12-04T09:38:23.1960624Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1961188Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1961192Z 2025-12-04T09:38:23.1961367Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1961993Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1962001Z 2025-12-04T09:38:23.1962214Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1967947Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1967961Z 2025-12-04T09:38:23.1972438Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1973104Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1973111Z 2025-12-04T09:38:23.1973356Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1974042Z FAILED [0.0010s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1974047Z 2025-12-04T09:38:23.1974241Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1974971Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1974996Z 2025-12-04T09:38:23.1975222Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1975792Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1975797Z 2025-12-04T09:38:23.1975980Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1976619Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1976624Z 2025-12-04T09:38:23.1976843Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1977414Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1977420Z 2025-12-04T09:38:23.1977601Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1978241Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1978246Z 2025-12-04T09:38:23.1978459Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1979026Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1979030Z 2025-12-04T09:38:23.1979208Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1979847Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1979857Z 2025-12-04T09:38:23.1980164Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1980737Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1980741Z 2025-12-04T09:38:23.1980924Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1981553Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1981557Z 2025-12-04T09:38:23.1981775Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1982345Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1982441Z 2025-12-04T09:38:23.1982629Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1983269Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1983274Z 2025-12-04T09:38:23.1983489Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1984060Z FAILED [0.0010s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1984065Z 2025-12-04T09:38:23.1984241Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1984873Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1984921Z 2025-12-04T09:38:23.1985140Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1985710Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1985719Z 2025-12-04T09:38:23.1985895Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1986534Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1986539Z 2025-12-04T09:38:23.1986759Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1987326Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1987334Z 2025-12-04T09:38:23.1987519Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1988155Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1988159Z 2025-12-04T09:38:23.1988377Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1988949Z FAILED [0.0008s] inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel - AttributeError: args 2025-12-04T09:38:23.1988954Z 2025-12-04T09:38:23.1989133Z To execute this test, run the following from the base repo dir: 2025-12-04T09:38:23.1989770Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_autograd.py TestAutogradWithCompiledAutograd.test_checkpointing_without_reentrant_dataparallel 2025-12-04T09:38:23.1989777Z 2025-12-04T09:38:23.1989994Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:38:23.1990141Z ========================= 49 failed, 1 passed in 9.25s ========================= 2025-12-04T09:38:23.1990226Z 2025-12-04T09:38:23.1990701Z FINISHED PRINTING LOG FILE of inductor/test_compiled_autograd 1/2 (test/test-reports/inductor.test_compiled_autograd_1.2_1749786211dd0efc_.log) 2025-12-04T09:38:23.1990707Z 2025-12-04T09:38:23.1991005Z Finished inductor/test_compiled_autograd 1/2 ... [2025-12-04 09:38:23.086935][1605.503406006], took 0.30min 2025-12-04T09:38:23.1991680Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_compiled_autograd/inductor.test_compiled_autograd-857de25fa6c3d7b5.xml 2025-12-04T09:38:23.4074800Z Uploading logs for 57118183167 to S3 2025-12-04T09:38:23.4449708Z Uploading artifacts took 0.28 seconds 2025-12-04T09:38:23.4450101Z inductor/test_compiled_autograd 1/2 failed! 2025-12-04T09:38:23.4454030Z Running test_testing 1/1 ... [2025-12-04 09:38:23.445096][1605.861568788] 2025-12-04T09:38:23.4454362Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:38:23.4458079Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_testing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:38:23.445458] 2025-12-04T09:38:32.0768513Z 2025-12-04T09:38:32.0769486Z test_testing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_testing_1.1_ee905ad7d74b3424_.log 2025-12-04T09:38:32.0800352Z Running 100 items in this shard: test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies 2025-12-04T09:38:32.0831414Z 2025-12-04T09:38:32.0831654Z Finished test_testing 1/1 ... [2025-12-04 09:38:32.076609][1614.493082077], took 0.14min 2025-12-04T09:38:32.0832512Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_testing/test_testing-509238380d0e9f60.xml 2025-12-04T09:38:32.1551689Z Running inductor/test_autoheuristic 1/1 ... [2025-12-04 09:38:32.154790][1614.571262241] 2025-12-04T09:38:32.1552205Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:38:32.1554936Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_autoheuristic.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:38:32.155105] 2025-12-04T09:38:38.3803124Z 2025-12-04T09:38:38.3804213Z inductor/test_autoheuristic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_autoheuristic_1.1_667ce2ef9a0c4c6d_.log 2025-12-04T09:38:38.3804997Z Running 0 items in this shard: 2025-12-04T09:38:38.3805204Z 2025-12-04T09:38:38.3805497Z Finished inductor/test_autoheuristic 1/1 ... [2025-12-04 09:38:38.379945][1620.796418005], took 0.10min 2025-12-04T09:38:38.3852551Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_autoheuristic/inductor.test_autoheuristic-b1e93791ac32a57e.xml 2025-12-04T09:38:38.4528421Z Running inductor/test_cutedsl_template 1/1 ... [2025-12-04 09:38:38.452416][1620.868888857] 2025-12-04T09:38:38.4529028Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:38:38.4531824Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cutedsl_template.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:38:38.452731] 2025-12-04T09:38:44.6790141Z 2025-12-04T09:38:44.6792131Z inductor/test_cutedsl_template 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cutedsl_template_1.1_764302bf470628f6_.log 2025-12-04T09:38:44.6793290Z Running 0 items in this shard: 2025-12-04T09:38:44.6793496Z 2025-12-04T09:38:44.6793802Z Finished inductor/test_cutedsl_template 1/1 ... [2025-12-04 09:38:44.678534][1627.095005415], took 0.10min 2025-12-04T09:38:44.6842841Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cutedsl_template/inductor.test_cutedsl_template-47a7a96144df1709.xml 2025-12-04T09:38:44.7552111Z Running inductor/test_benchmark_fusion 1/1 ... [2025-12-04 09:38:44.754852][1627.171324969] 2025-12-04T09:38:44.7552704Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:38:44.7555673Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_benchmark_fusion.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:38:44.755201] 2025-12-04T09:41:12.8804062Z 2025-12-04T09:41:12.8808666Z PRINTING LOG FILE of inductor/test_benchmark_fusion 1/1 (test/test-reports/inductor.test_benchmark_fusion_1.1_baf135911d2d356d_.log) 2025-12-04T09:41:12.8809850Z Test results will be stored in test-reports/python-pytest/inductor.test_benchmark_fusion/inductor.test_benchmark_fusion-1a5c9a2f1b3021da.xml 2025-12-04T09:41:12.8810673Z ============================= test session starts ============================== 2025-12-04T09:41:12.8811229Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:41:12.8811890Z cachedir: .pytest_cache 2025-12-04T09:41:12.8812697Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:41:12.8815110Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:41:12.8815524Z configfile: pytest.ini 2025-12-04T09:41:12.8816387Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:41:12.8817247Z collecting ... collected 16 items 2025-12-04T09:41:12.8817736Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:41:12.8870214Z Running 100 items in this shard: test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code, test/inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code 2025-12-04T09:41:12.8914066Z 2025-12-04T09:41:12.8914980Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 1%] 2025-12-04T09:41:12.8916455Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [5.2188s] [ 2%] 2025-12-04T09:41:12.8917908Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8919888Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8921745Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8923611Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8925516Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8927456Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8929322Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0008s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8931185Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8933056Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8934909Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8937250Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8939295Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8941637Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8944256Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8946114Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8948093Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8949947Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8951811Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8953661Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8955592Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8957445Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8959297Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8961172Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8963034Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8964880Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8966737Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8968776Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8970639Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8972494Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8974362Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8976273Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0008s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8978130Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8979990Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8981902Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8983755Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8985613Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8987467Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8989339Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8991242Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8993111Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8995059Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8996916Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.8998777Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.9000630Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.9002575Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.9004645Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.9006671Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.9008875Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.9010834Z inductor/test_benchmark_fusion.py::BenchmarkFusionGpuTest::test_tield_kernel_fusion_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:41:12.9012289Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.5568s] [ 2%] 2025-12-04T09:41:12.9013647Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code W1204 09:39:02.839000 12194 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpqun5dtux 2025-12-04T09:41:12.9014648Z FAILED [4.3923s] [ 2%] 2025-12-04T09:41:12.9015257Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3646s] [ 2%] 2025-12-04T09:41:12.9016627Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code W1204 09:39:09.592000 12194 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp2w3f9rlp 2025-12-04T09:41:12.9017619Z FAILED [4.3588s] [ 2%] 2025-12-04T09:41:12.9018217Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4389s] [ 2%] 2025-12-04T09:41:12.9019223Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.6287s] [ 2%] 2025-12-04T09:41:12.9020226Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.7775s] [ 2%] 2025-12-04T09:41:12.9021219Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.9777s] [ 2%] 2025-12-04T09:41:12.9022225Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3659s] [ 2%] 2025-12-04T09:41:12.9023331Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3605s] [ 2%] 2025-12-04T09:41:12.9024338Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4610s] [ 2%] 2025-12-04T09:41:12.9025338Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.7598s] [ 2%] 2025-12-04T09:41:12.9026332Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.2953s] [ 2%] 2025-12-04T09:41:12.9027684Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code W1204 09:39:36.591000 12194 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmphd5usaet 2025-12-04T09:41:12.9028723Z FAILED [3.9186s] [ 2%] 2025-12-04T09:41:12.9029681Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code W1204 09:39:40.252000 12194 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpc9pvlbau 2025-12-04T09:41:12.9030676Z FAILED [3.6587s] [ 2%] 2025-12-04T09:41:12.9048053Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4134s] [ 2%] 2025-12-04T09:41:12.9049073Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.6895s] [ 2%] 2025-12-04T09:41:12.9050430Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code W1204 09:39:49.396000 12194 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpg8i5cgkd 2025-12-04T09:41:12.9051427Z FAILED [4.0362s] [ 2%] 2025-12-04T09:41:12.9052173Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3042s] [ 2%] 2025-12-04T09:41:12.9053192Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.5841s] [ 2%] 2025-12-04T09:41:12.9054190Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [3.0050s] [ 2%] 2025-12-04T09:41:12.9055195Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.7041s] [ 2%] 2025-12-04T09:41:12.9056205Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3384s] [ 2%] 2025-12-04T09:41:12.9057211Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.5317s] [ 2%] 2025-12-04T09:41:12.9058219Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.5557s] [ 2%] 2025-12-04T09:41:12.9059224Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3038s] [ 2%] 2025-12-04T09:41:12.9060229Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.2737s] [ 2%] 2025-12-04T09:41:12.9061243Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4533s] [ 2%] 2025-12-04T09:41:12.9062252Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3014s] [ 2%] 2025-12-04T09:41:12.9063250Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4182s] [ 2%] 2025-12-04T09:41:12.9064262Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [3.0054s] [ 2%] 2025-12-04T09:41:12.9065633Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code W1204 09:40:26.423000 12194 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpzrw3psn5 2025-12-04T09:41:12.9066936Z FAILED [4.2237s] [ 2%] 2025-12-04T09:41:12.9068062Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code W1204 09:40:30.721000 12194 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpqepf3qkx 2025-12-04T09:41:12.9069237Z FAILED [4.2954s] [ 2%] 2025-12-04T09:41:12.9069957Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.6373s] [ 2%] 2025-12-04T09:41:12.9071160Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.9159s] [ 2%] 2025-12-04T09:41:12.9072329Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3026s] [ 2%] 2025-12-04T09:41:12.9073507Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4377s] [ 2%] 2025-12-04T09:41:12.9074760Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.2997s] [ 2%] 2025-12-04T09:41:12.9075935Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3009s] [ 2%] 2025-12-04T09:41:12.9077101Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4198s] [ 2%] 2025-12-04T09:41:12.9078278Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.5923s] [ 2%] 2025-12-04T09:41:12.9079454Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.6731s] [ 2%] 2025-12-04T09:41:12.9080681Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3323s] [ 2%] 2025-12-04T09:41:12.9081904Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3762s] [ 2%] 2025-12-04T09:41:12.9083079Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4582s] [ 2%] 2025-12-04T09:41:12.9084253Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3603s] [ 2%] 2025-12-04T09:41:12.9085428Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [3.0208s] [ 2%] 2025-12-04T09:41:12.9086417Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.3134s] [ 2%] 2025-12-04T09:41:12.9087490Z inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code PASSED [2.4347s] [ 2%] 2025-12-04T09:41:12.9088056Z 2025-12-04T09:41:12.9088187Z =================================== FAILURES =================================== 2025-12-04T09:41:12.9088709Z ______ BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code _______ 2025-12-04T09:41:12.9089196Z Traceback (most recent call last): 2025-12-04T09:41:12.9089813Z File "/var/lib/jenkins/workspace/test/inductor/test_benchmark_fusion.py", line 303, in test_equivalent_template_code 2025-12-04T09:41:12.9090509Z ).check("" if config.cpp_wrapper else "return").run(out_code[0]) 2025-12-04T09:41:12.9091048Z RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:12.9091494Z Searched string: 2025-12-04T09:41:12.9091745Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:12.9092056Z torch.cuda.set_device(0) 2025-12-04T09:41:12.9092416Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:12.9092941Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:12.9093393Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9093798Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:12.9094297Z del arg0_1 2025-12-04T09:41:12.9094527Z del arg2_1 2025-12-04T09:41:12.9094776Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:12.9095234Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:12.9095704Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9096106Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:12.9096494Z del arg1_1 2025-12-04T09:41:12.9096731Z return (buf1, ) 2025-12-04T09:41:12.9096879Z 2025-12-04T09:41:12.9096992Z runner = Runner(partitions=[]) 2025-12-04T09:41:12.9097255Z call = runner.call 2025-12-04T09:41:12.9097561Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:12.9097830Z 2025-12-04T09:41:12.9097885Z 2025-12-04T09:41:12.9098030Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:12.9098426Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:12.9098812Z from torch._inductor.utils import print_performance 2025-12-04T09:41:12.9099275Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9099786Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9100288Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9100724Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:12.9101113Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:12.9101386Z 2025-12-04T09:41:12.9101390Z 2025-12-04T09:41:12.9101491Z if __name__ == "__main__": 2025-12-04T09:41:12.9101852Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:12.9102369Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:12.9102755Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:12.9102969Z 2025-12-04T09:41:12.9102979Z 2025-12-04T09:41:12.9103166Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:12.9104106Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:12.9104865Z 2025-12-04T09:41:12.9105085Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:12.9105595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9105982Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9106290Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9108217Z inductor [('triton_bundler_save_kernel', 216), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('async_compile_cache_miss', 7), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9110249Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9110692Z graph_break [] 2025-12-04T09:41:12.9110922Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9111309Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9112917Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py:3433: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage() 2025-12-04T09:41:12.9114521Z current_size = base.storage().size() 2025-12-04T09:41:12.9114816Z Autotune Choices Stats: 2025-12-04T09:41:12.9115807Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_10", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9116831Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9117095Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9117361Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9118045Z triton_mm_10 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9119161Z triton_mm_9 0.0277 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9120207Z triton_mm_11 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9121242Z triton_mm_13 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9122300Z triton_mm_14 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9123357Z triton_mm_12 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:12.9124444Z triton_mm_1 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9125472Z triton_mm_2 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9126486Z triton_mm_6 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9127602Z triton_mm_7 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9128524Z SingleProcess AUTOTUNE benchmarking takes 0.1729 seconds and 1.0713 seconds precompiling for 15 choices 2025-12-04T09:41:12.9129350Z Compiled module path: /tmp/tmpgejy12ej/qv/cqvwpzgbltbzt3unpdzscv64a2elpufsmhey6btnq73pj2f4vaa7.py 2025-12-04T09:41:12.9130185Z Compiled module path: /tmp/tmpgejy12ej/jj/cjj4tsxlplxvzz5g4jfphbz5jsxl7b3fd5mxbhxre5lcbuh3vmrc.py 2025-12-04T09:41:12.9130826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9131216Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9131540Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9132024Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9134090Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9136382Z graph_break [] 2025-12-04T09:41:12.9136837Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9137234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9137615Z Autotune Choices Stats: 2025-12-04T09:41:12.9138603Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_30", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9139624Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9139891Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9140166Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9140851Z triton_mm_30 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9141968Z triton_mm_32 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9143012Z triton_mm_33 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9144050Z triton_mm_34 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9145092Z triton_mm_35 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9146186Z triton_mm_36 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9147236Z triton_mm_37 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9148274Z triton_mm_38 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9149325Z triton_mm_39 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9150432Z triton_mm_40 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9151350Z SingleProcess AUTOTUNE benchmarking takes 0.2013 seconds and 0.7407 seconds precompiling for 15 choices 2025-12-04T09:41:12.9152162Z Compiled module path: /tmp/tmpu9odmkxw/ai/cai5yet7zvfv3ry4tqny7aptct3cbabbgknxxt3otvygyekzraji.py 2025-12-04T09:41:12.9152975Z Compiled module path: /tmp/tmpu9odmkxw/nq/cnqm37sxiqmqocnfb7wdcdqxnd56ra26fhpzusixi763umecdh56.py 2025-12-04T09:41:12.9153614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9153996Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9154308Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9154796Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9157004Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 36), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 21), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 14), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9158955Z graph_break [] 2025-12-04T09:41:12.9159195Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9159574Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9159959Z Autotune Choices Stats: 2025-12-04T09:41:12.9160947Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_73", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9161974Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9162234Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9162586Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9163285Z triton_mm_73 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9164332Z triton_mm_62 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9165360Z triton_mm_65 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9166394Z triton_mm_71 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9167544Z triton_mm_61 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9168579Z triton_mm_63 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9169596Z triton_mm_60 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9170616Z triton_mm_64 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9171645Z triton_mm_67 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9172698Z triton_mm_68 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9173616Z SingleProcess AUTOTUNE benchmarking takes 0.2006 seconds and 0.6267 seconds precompiling for 15 choices 2025-12-04T09:41:12.9174415Z Compiled module path: /tmp/tmpqun5dtux/q7/cq7yaif2wknvncdje2hbdwcngroryoj2mtkk4dkwdkx45aat6hm7.py 2025-12-04T09:41:12.9175234Z Compiled module path: /tmp/tmpqun5dtux/y6/cy66mazx6rmemwkg6jecmgbfmpsvz6cfz5y45krjnl3azaph6y7t.py 2025-12-04T09:41:12.9175975Z ______ BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code _______ 2025-12-04T09:41:12.9176472Z Traceback (most recent call last): 2025-12-04T09:41:12.9177078Z File "/var/lib/jenkins/workspace/test/inductor/test_benchmark_fusion.py", line 303, in test_equivalent_template_code 2025-12-04T09:41:12.9177783Z ).check("" if config.cpp_wrapper else "return").run(out_code[0]) 2025-12-04T09:41:12.9178325Z RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:12.9178772Z Searched string: 2025-12-04T09:41:12.9179027Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:12.9179425Z torch.cuda.set_device(0) 2025-12-04T09:41:12.9179789Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:12.9180315Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:12.9180767Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9181186Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:12.9181589Z del arg0_1 2025-12-04T09:41:12.9181820Z del arg2_1 2025-12-04T09:41:12.9182068Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:12.9182532Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:12.9183008Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9183410Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:12.9183853Z del arg1_1 2025-12-04T09:41:12.9184080Z return (buf1, ) 2025-12-04T09:41:12.9184245Z 2025-12-04T09:41:12.9184354Z runner = Runner(partitions=[]) 2025-12-04T09:41:12.9184622Z call = runner.call 2025-12-04T09:41:12.9184920Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:12.9185186Z 2025-12-04T09:41:12.9185190Z 2025-12-04T09:41:12.9185335Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:12.9185720Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:12.9186105Z from torch._inductor.utils import print_performance 2025-12-04T09:41:12.9186565Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9187070Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9187615Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9188059Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:12.9188446Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:12.9188722Z 2025-12-04T09:41:12.9188726Z 2025-12-04T09:41:12.9188824Z if __name__ == "__main__": 2025-12-04T09:41:12.9189199Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:12.9189669Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:12.9190052Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:12.9190269Z 2025-12-04T09:41:12.9190273Z 2025-12-04T09:41:12.9190466Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:12.9191392Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:12.9192150Z 2025-12-04T09:41:12.9192371Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:12.9192878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9193268Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9193578Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9195497Z inductor [('triton_bundler_save_kernel', 216), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('async_compile_cache_miss', 7), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9197514Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9197965Z graph_break [] 2025-12-04T09:41:12.9198201Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9198580Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9200265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py:3433: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage() 2025-12-04T09:41:12.9201772Z current_size = base.storage().size() 2025-12-04T09:41:12.9202068Z Autotune Choices Stats: 2025-12-04T09:41:12.9203046Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_10", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9204106Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9204365Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9204642Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9205318Z triton_mm_10 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9206357Z triton_mm_9 0.0277 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9207452Z triton_mm_11 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9208489Z triton_mm_13 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9209591Z triton_mm_14 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9210633Z triton_mm_12 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:12.9211671Z triton_mm_1 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9212691Z triton_mm_2 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9213705Z triton_mm_6 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9214735Z triton_mm_7 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9215644Z SingleProcess AUTOTUNE benchmarking takes 0.1729 seconds and 1.0713 seconds precompiling for 15 choices 2025-12-04T09:41:12.9216446Z Compiled module path: /tmp/tmpgejy12ej/qv/cqvwpzgbltbzt3unpdzscv64a2elpufsmhey6btnq73pj2f4vaa7.py 2025-12-04T09:41:12.9217265Z Compiled module path: /tmp/tmpgejy12ej/jj/cjj4tsxlplxvzz5g4jfphbz5jsxl7b3fd5mxbhxre5lcbuh3vmrc.py 2025-12-04T09:41:12.9217898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9218279Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9218595Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9219081Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9221357Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9223210Z graph_break [] 2025-12-04T09:41:12.9223446Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9223828Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9224198Z Autotune Choices Stats: 2025-12-04T09:41:12.9225182Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_30", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9226243Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9226507Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9226777Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9227460Z triton_mm_30 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9228504Z triton_mm_32 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9229535Z triton_mm_33 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9230602Z triton_mm_34 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9231642Z triton_mm_35 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9232672Z triton_mm_36 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9233708Z triton_mm_37 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9234760Z triton_mm_38 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9236103Z triton_mm_39 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9237164Z triton_mm_40 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9238088Z SingleProcess AUTOTUNE benchmarking takes 0.2013 seconds and 0.7407 seconds precompiling for 15 choices 2025-12-04T09:41:12.9238903Z Compiled module path: /tmp/tmpu9odmkxw/ai/cai5yet7zvfv3ry4tqny7aptct3cbabbgknxxt3otvygyekzraji.py 2025-12-04T09:41:12.9239723Z Compiled module path: /tmp/tmpu9odmkxw/nq/cnqm37sxiqmqocnfb7wdcdqxnd56ra26fhpzusixi763umecdh56.py 2025-12-04T09:41:12.9240374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9240765Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9241100Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9241592Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9243920Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 36), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 21), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 14), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9245890Z graph_break [] 2025-12-04T09:41:12.9246145Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9246525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9246921Z Autotune Choices Stats: 2025-12-04T09:41:12.9248273Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_73", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9249495Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9249765Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9250048Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9250845Z triton_mm_73 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9252101Z triton_mm_62 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9253388Z triton_mm_65 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9254633Z triton_mm_71 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9255862Z triton_mm_61 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9257082Z triton_mm_63 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9258299Z triton_mm_60 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9259508Z triton_mm_64 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9260738Z triton_mm_67 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9261971Z triton_mm_68 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9263038Z SingleProcess AUTOTUNE benchmarking takes 0.2006 seconds and 0.6267 seconds precompiling for 15 choices 2025-12-04T09:41:12.9263959Z Compiled module path: /tmp/tmpqun5dtux/q7/cq7yaif2wknvncdje2hbdwcngroryoj2mtkk4dkwdkx45aat6hm7.py 2025-12-04T09:41:12.9264921Z Compiled module path: /tmp/tmpqun5dtux/y6/cy66mazx6rmemwkg6jecmgbfmpsvz6cfz5y45krjnl3azaph6y7t.py 2025-12-04T09:41:12.9265672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9266111Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9266448Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9267089Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9269583Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9271896Z graph_break [] 2025-12-04T09:41:12.9272138Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9272565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9273027Z Autotune Choices Stats: 2025-12-04T09:41:12.9274234Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_91", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9275405Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9275687Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9275974Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9276742Z triton_mm_91 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9277935Z triton_mm_95 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9279142Z triton_mm_90 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9280268Z triton_mm_92 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9281438Z triton_mm_93 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9282542Z triton_mm_96 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9283593Z triton_mm_97 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9284665Z triton_mm_98 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9285708Z triton_mm_94 0.0307 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9286754Z triton_mm_100 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9287755Z SingleProcess AUTOTUNE benchmarking takes 0.2153 seconds and 0.6426 seconds precompiling for 15 choices 2025-12-04T09:41:12.9288571Z Compiled module path: /tmp/tmp2v_kzp9a/fj/cfjz3b7ajfatqwomwryerntjnvk6kd5wysrcdign42gjpmvqkh3t.py 2025-12-04T09:41:12.9289375Z Compiled module path: /tmp/tmp2v_kzp9a/57/c57neznqb6ba2jtzlgbchzekwlmukri7wpo65rmsqusn7t6xlfxx.py 2025-12-04T09:41:12.9290013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9290501Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9290830Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9291325Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9293486Z inductor [('triton_bundler_save_kernel', 288), ('benchmarking.InductorBenchmarker.benchmark_gpu', 38), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 23), ('coordesc_tuning_bench', 16), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9295508Z graph_break [] 2025-12-04T09:41:12.9295744Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9296145Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9296530Z Autotune Choices Stats: 2025-12-04T09:41:12.9297532Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_134", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9298581Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9298847Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9299125Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9299812Z triton_mm_134 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9300961Z triton_mm_121 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9302014Z triton_mm_122 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9303058Z triton_mm_124 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9304094Z triton_mm_125 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9305136Z triton_mm_128 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9306192Z triton_mm_129 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9307243Z triton_mm_130 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9308286Z triton_mm_131 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9309327Z triton_mm_120 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9310281Z SingleProcess AUTOTUNE benchmarking takes 0.2007 seconds and 0.6630 seconds precompiling for 15 choices 2025-12-04T09:41:12.9311085Z Compiled module path: /tmp/tmp2w3f9rlp/jd/cjdbzzfn2s4jgc6gkasctk4suc4halinn3hncy6u26d22lzxdphd.py 2025-12-04T09:41:12.9311997Z Compiled module path: /tmp/tmp2w3f9rlp/4s/c4sivzup5qddnp7nobqsap2nr2ndjcdy3tashkjs4q3gubdqirss.py 2025-12-04T09:41:12.9312739Z ______ BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code _______ 2025-12-04T09:41:12.9313240Z Traceback (most recent call last): 2025-12-04T09:41:12.9313867Z File "/var/lib/jenkins/workspace/test/inductor/test_benchmark_fusion.py", line 303, in test_equivalent_template_code 2025-12-04T09:41:12.9314574Z ).check("" if config.cpp_wrapper else "return").run(out_code[0]) 2025-12-04T09:41:12.9315115Z RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:12.9315571Z Searched string: 2025-12-04T09:41:12.9315832Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:12.9316151Z torch.cuda.set_device(0) 2025-12-04T09:41:12.9316533Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:12.9317114Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:12.9317581Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9317998Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:12.9318413Z del arg0_1 2025-12-04T09:41:12.9318657Z del arg2_1 2025-12-04T09:41:12.9318901Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:12.9319381Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:12.9319858Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9320259Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:12.9320664Z del arg1_1 2025-12-04T09:41:12.9320900Z return (buf1, ) 2025-12-04T09:41:12.9321054Z 2025-12-04T09:41:12.9321211Z runner = Runner(partitions=[]) 2025-12-04T09:41:12.9321480Z call = runner.call 2025-12-04T09:41:12.9321792Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:12.9322053Z 2025-12-04T09:41:12.9322057Z 2025-12-04T09:41:12.9322215Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:12.9322597Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:12.9322989Z from torch._inductor.utils import print_performance 2025-12-04T09:41:12.9323452Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9323973Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9324478Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9324913Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:12.9325307Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:12.9325583Z 2025-12-04T09:41:12.9325587Z 2025-12-04T09:41:12.9325688Z if __name__ == "__main__": 2025-12-04T09:41:12.9326056Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:12.9326547Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:12.9326935Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:12.9327238Z 2025-12-04T09:41:12.9327242Z 2025-12-04T09:41:12.9327424Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:12.9328370Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:12.9329135Z 2025-12-04T09:41:12.9329355Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:12.9329870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9330251Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9330577Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9332599Z inductor [('triton_bundler_save_kernel', 216), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('async_compile_cache_miss', 7), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9334649Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9335090Z graph_break [] 2025-12-04T09:41:12.9335663Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9354859Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9356606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py:3433: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage() 2025-12-04T09:41:12.9358298Z current_size = base.storage().size() 2025-12-04T09:41:12.9358593Z Autotune Choices Stats: 2025-12-04T09:41:12.9359596Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_10", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9360631Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9360899Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9361167Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9361928Z triton_mm_10 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9362988Z triton_mm_9 0.0277 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9364036Z triton_mm_11 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9365071Z triton_mm_13 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9366121Z triton_mm_14 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9367286Z triton_mm_12 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:12.9368328Z triton_mm_1 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9369349Z triton_mm_2 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9370365Z triton_mm_6 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9371403Z triton_mm_7 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9372322Z SingleProcess AUTOTUNE benchmarking takes 0.1729 seconds and 1.0713 seconds precompiling for 15 choices 2025-12-04T09:41:12.9373363Z Compiled module path: /tmp/tmpgejy12ej/qv/cqvwpzgbltbzt3unpdzscv64a2elpufsmhey6btnq73pj2f4vaa7.py 2025-12-04T09:41:12.9374184Z Compiled module path: /tmp/tmpgejy12ej/jj/cjj4tsxlplxvzz5g4jfphbz5jsxl7b3fd5mxbhxre5lcbuh3vmrc.py 2025-12-04T09:41:12.9374825Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9375216Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9375530Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9376022Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9378079Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9379994Z graph_break [] 2025-12-04T09:41:12.9380240Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9380619Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9381005Z Autotune Choices Stats: 2025-12-04T09:41:12.9381988Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_30", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9383012Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9383313Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9383591Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9384277Z triton_mm_30 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9385314Z triton_mm_32 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9386361Z triton_mm_33 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9387400Z triton_mm_34 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9388435Z triton_mm_35 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9389477Z triton_mm_36 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9390556Z triton_mm_37 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9391604Z triton_mm_38 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9392656Z triton_mm_39 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9393704Z triton_mm_40 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9394701Z SingleProcess AUTOTUNE benchmarking takes 0.2013 seconds and 0.7407 seconds precompiling for 15 choices 2025-12-04T09:41:12.9395497Z Compiled module path: /tmp/tmpu9odmkxw/ai/cai5yet7zvfv3ry4tqny7aptct3cbabbgknxxt3otvygyekzraji.py 2025-12-04T09:41:12.9396326Z Compiled module path: /tmp/tmpu9odmkxw/nq/cnqm37sxiqmqocnfb7wdcdqxnd56ra26fhpzusixi763umecdh56.py 2025-12-04T09:41:12.9396973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9397366Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9397684Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9398176Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9400356Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 36), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 21), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 14), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9402349Z graph_break [] 2025-12-04T09:41:12.9402582Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9402969Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9403359Z Autotune Choices Stats: 2025-12-04T09:41:12.9404354Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_73", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9405413Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9405676Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9405961Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9406641Z triton_mm_73 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9407740Z triton_mm_62 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9408769Z triton_mm_65 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9409800Z triton_mm_71 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9410854Z triton_mm_61 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9411879Z triton_mm_63 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9412903Z triton_mm_60 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9413937Z triton_mm_64 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9414967Z triton_mm_67 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9416105Z triton_mm_68 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9417009Z SingleProcess AUTOTUNE benchmarking takes 0.2006 seconds and 0.6267 seconds precompiling for 15 choices 2025-12-04T09:41:12.9417810Z Compiled module path: /tmp/tmpqun5dtux/q7/cq7yaif2wknvncdje2hbdwcngroryoj2mtkk4dkwdkx45aat6hm7.py 2025-12-04T09:41:12.9418637Z Compiled module path: /tmp/tmpqun5dtux/y6/cy66mazx6rmemwkg6jecmgbfmpsvz6cfz5y45krjnl3azaph6y7t.py 2025-12-04T09:41:12.9419281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9419656Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9419971Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9420462Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9422549Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9424402Z graph_break [] 2025-12-04T09:41:12.9424642Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9425033Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9425417Z Autotune Choices Stats: 2025-12-04T09:41:12.9426396Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_91", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9427456Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9427724Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9428003Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9428672Z triton_mm_91 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9429726Z triton_mm_95 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9430757Z triton_mm_90 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9431787Z triton_mm_92 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9432811Z triton_mm_93 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9433837Z triton_mm_96 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9434862Z triton_mm_97 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9436203Z triton_mm_98 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9437249Z triton_mm_94 0.0307 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9438418Z triton_mm_100 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9439351Z SingleProcess AUTOTUNE benchmarking takes 0.2153 seconds and 0.6426 seconds precompiling for 15 choices 2025-12-04T09:41:12.9440155Z Compiled module path: /tmp/tmp2v_kzp9a/fj/cfjz3b7ajfatqwomwryerntjnvk6kd5wysrcdign42gjpmvqkh3t.py 2025-12-04T09:41:12.9440965Z Compiled module path: /tmp/tmp2v_kzp9a/57/c57neznqb6ba2jtzlgbchzekwlmukri7wpo65rmsqusn7t6xlfxx.py 2025-12-04T09:41:12.9441589Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9441993Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9442323Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9442875Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9445019Z inductor [('triton_bundler_save_kernel', 288), ('benchmarking.InductorBenchmarker.benchmark_gpu', 38), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 23), ('coordesc_tuning_bench', 16), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9447019Z graph_break [] 2025-12-04T09:41:12.9447268Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9447672Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9448110Z Autotune Choices Stats: 2025-12-04T09:41:12.9449134Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_134", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9450172Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9450442Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9450707Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9451401Z triton_mm_134 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9452469Z triton_mm_121 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9453520Z triton_mm_122 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9454579Z triton_mm_124 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9455619Z triton_mm_125 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9456666Z triton_mm_128 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9457711Z triton_mm_129 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9458755Z triton_mm_130 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9459908Z triton_mm_131 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9460945Z triton_mm_120 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9461858Z SingleProcess AUTOTUNE benchmarking takes 0.2007 seconds and 0.6630 seconds precompiling for 15 choices 2025-12-04T09:41:12.9462658Z Compiled module path: /tmp/tmp2w3f9rlp/jd/cjdbzzfn2s4jgc6gkasctk4suc4halinn3hncy6u26d22lzxdphd.py 2025-12-04T09:41:12.9463466Z Compiled module path: /tmp/tmp2w3f9rlp/4s/c4sivzup5qddnp7nobqsap2nr2ndjcdy3tashkjs4q3gubdqirss.py 2025-12-04T09:41:12.9464107Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9464528Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9464851Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9465341Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9467380Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9469239Z graph_break [] 2025-12-04T09:41:12.9469484Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9469900Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9470279Z Autotune Choices Stats: 2025-12-04T09:41:12.9471266Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_158", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9472280Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9472536Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9472809Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9473497Z triton_mm_158 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9474552Z triton_mm_159 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9475613Z triton_mm_161 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9476671Z triton_mm_150 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9477712Z triton_mm_151 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9478749Z triton_mm_153 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9479781Z triton_mm_154 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9480892Z triton_mm_155 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9481924Z triton_mm_156 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9482960Z triton_mm_157 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9483874Z SingleProcess AUTOTUNE benchmarking takes 0.2054 seconds and 0.6100 seconds precompiling for 15 choices 2025-12-04T09:41:12.9484674Z Compiled module path: /tmp/tmpwi6t5aeb/gn/cgneadkz2tnedarmlzhqm4gn55whqeexaln2fzmkic7oxhyres4r.py 2025-12-04T09:41:12.9485497Z Compiled module path: /tmp/tmpwi6t5aeb/gy/cgyxpcavcgmmp26t6h56dechg7qqid3n7dyu5q6u5h3b6c7fepoj.py 2025-12-04T09:41:12.9486187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9486580Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9486893Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9487457Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9489495Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9491405Z graph_break [] 2025-12-04T09:41:12.9491638Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9492033Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9492416Z Autotune Choices Stats: 2025-12-04T09:41:12.9493395Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_186", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9494400Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9494664Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9494946Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9495631Z triton_mm_186 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9496675Z triton_mm_193 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9497737Z triton_mm_182 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9498781Z triton_mm_194 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9499823Z triton_mm_184 0.0308 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9500857Z triton_mm_187 0.0317 ms 90.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9501970Z triton_mm_185 0.0317 ms 90.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9503007Z triton_mm_180 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9504049Z triton_mm_181 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9505080Z triton_mm_183 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9505980Z SingleProcess AUTOTUNE benchmarking takes 0.2087 seconds and 0.6406 seconds precompiling for 15 choices 2025-12-04T09:41:12.9506779Z Compiled module path: /tmp/tmpk78l5vwh/kk/ckk35oaicdyefn5yev5gn7hgiqa5qquexctvfdcqebmw6q6whm6z.py 2025-12-04T09:41:12.9507646Z Compiled module path: /tmp/tmpk78l5vwh/hm/chmuaf5ovlicrcx6s7zj4pujz62mn7yxmmkafqcdbwqtfoym5ys5.py 2025-12-04T09:41:12.9508290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9508669Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9508992Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9509505Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9511804Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9513904Z graph_break [] 2025-12-04T09:41:12.9514195Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9514594Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9514976Z Autotune Choices Stats: 2025-12-04T09:41:12.9515962Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_220", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9516989Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9517260Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9517529Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9518212Z triton_mm_220 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9519281Z triton_mm_211 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9520324Z triton_mm_212 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9521359Z triton_mm_214 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9522391Z triton_mm_215 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9523415Z triton_mm_216 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9524564Z triton_mm_217 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9525618Z triton_mm_218 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9526673Z triton_mm_219 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9527820Z triton_mm_221 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9528746Z SingleProcess AUTOTUNE benchmarking takes 0.1998 seconds and 0.6553 seconds precompiling for 15 choices 2025-12-04T09:41:12.9529622Z Compiled module path: /tmp/tmpwuxhdg2k/ip/cipioi7ficmn3ww6fbrd2oyul2gc6oxnifncerak7mzlxvpbihdc.py 2025-12-04T09:41:12.9530459Z Compiled module path: /tmp/tmpwuxhdg2k/tq/ctqm7bgwweclnxyitnuulxziaydok5e5vcjg4prplje6cyryh3p3.py 2025-12-04T09:41:12.9531102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9531486Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9531802Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9532288Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9534330Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9536503Z graph_break [] 2025-12-04T09:41:12.9536743Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9537128Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9537499Z Autotune Choices Stats: 2025-12-04T09:41:12.9538492Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_250", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9539515Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9539785Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9540049Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9540745Z triton_mm_250 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9541794Z triton_mm_245 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9542837Z triton_mm_249 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9543895Z triton_mm_251 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9544930Z triton_mm_243 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9546102Z triton_mm_241 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9547135Z triton_mm_240 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9548172Z triton_mm_242 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9549202Z triton_mm_244 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9550239Z triton_mm_246 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9551226Z SingleProcess AUTOTUNE benchmarking takes 0.2009 seconds and 0.6388 seconds precompiling for 15 choices 2025-12-04T09:41:12.9552034Z Compiled module path: /tmp/tmpv4covkzm/xp/cxpol2uuscxtvrr3qpm7ebdm7l5updopjmnk75hpevamiwzj65o7.py 2025-12-04T09:41:12.9552851Z Compiled module path: /tmp/tmpv4covkzm/rn/crnctdx35xmvga3ti6xdrde3d3i7bt47oqtpexetsy437yfgpmyy.py 2025-12-04T09:41:12.9553479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9553867Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9554190Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9554679Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9556727Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9558656Z graph_break [] 2025-12-04T09:41:12.9558893Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9559279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9559662Z Autotune Choices Stats: 2025-12-04T09:41:12.9560647Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_276", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.02768000029027462, "best_triton_pos": 0} 2025-12-04T09:41:12.9561665Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9561927Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9562201Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9562883Z triton_mm_276 0.0277 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9563931Z triton_mm_272 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9564960Z triton_mm_273 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9566009Z triton_mm_277 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9567106Z triton_mm_280 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9568240Z triton_mm_281 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9569292Z triton_mm_278 0.0296 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9570318Z triton_mm_270 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9571358Z triton_mm_271 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9572440Z triton_mm_274 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9573343Z SingleProcess AUTOTUNE benchmarking takes 0.2011 seconds and 0.6415 seconds precompiling for 15 choices 2025-12-04T09:41:12.9574141Z Compiled module path: /tmp/tmpuyrn9hru/66/c66x747yoejfwu2glewx7h3jab2qxz2k5kaaujahxklodqyj4cnu.py 2025-12-04T09:41:12.9574941Z Compiled module path: /tmp/tmpuyrn9hru/y7/cy7cdv3w3mgfm2t6nqc7yr2wtagkyvkt74hbgdknhwoqeeslovzq.py 2025-12-04T09:41:12.9575574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9575958Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9576269Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9576761Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9578855Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9580716Z graph_break [] 2025-12-04T09:41:12.9580955Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9581331Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9581713Z Autotune Choices Stats: 2025-12-04T09:41:12.9582698Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_302", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9583728Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9583991Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9584261Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9584939Z triton_mm_302 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9585976Z triton_mm_303 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9587024Z triton_mm_305 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9588074Z triton_mm_307 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9589227Z triton_mm_309 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9590268Z triton_mm_300 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9591296Z triton_mm_301 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9592328Z triton_mm_304 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9593367Z triton_mm_306 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9594453Z triton_mm_308 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9595364Z SingleProcess AUTOTUNE benchmarking takes 0.2012 seconds and 0.6271 seconds precompiling for 15 choices 2025-12-04T09:41:12.9596178Z Compiled module path: /tmp/tmpnxfvglo0/mv/cmv5svnrqdgxfpijsujfb6lt3msltpw2rqji7ouqvhdeeykjqzmz.py 2025-12-04T09:41:12.9597008Z Compiled module path: /tmp/tmpnxfvglo0/rx/crxtbd5ejbufzmuyijmfl3i77mx7bc25cjvtl52cepcxot7477r3.py 2025-12-04T09:41:12.9597650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9598038Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9598361Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9598907Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9600961Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9602816Z graph_break [] 2025-12-04T09:41:12.9603055Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9603439Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9603828Z Autotune Choices Stats: 2025-12-04T09:41:12.9604838Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_333", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9605851Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9606128Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9606407Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9607152Z triton_mm_333 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9608214Z triton_mm_335 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9609289Z triton_mm_343 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9610504Z triton_mm_337 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9611562Z triton_mm_332 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9612605Z triton_mm_338 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9613672Z triton_mm_330 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9614719Z triton_mm_331 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9615839Z triton_mm_334 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9616871Z triton_mm_336 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9617769Z SingleProcess AUTOTUNE benchmarking takes 0.4010 seconds and 0.6248 seconds precompiling for 15 choices 2025-12-04T09:41:12.9618582Z Compiled module path: /tmp/tmp9c3250tk/oj/cojlvuknqajncjwhcs6yfkfjjtihtyenckj7m5l37blt326fq55c.py 2025-12-04T09:41:12.9619397Z Compiled module path: /tmp/tmp9c3250tk/st/cst5rwasvqebnwrew6gvackeisqs7qwte54od7766agjws6s3ura.py 2025-12-04T09:41:12.9620090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9620516Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9620841Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9621339Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9623379Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9625228Z graph_break [] 2025-12-04T09:41:12.9625481Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9625879Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9626270Z Autotune Choices Stats: 2025-12-04T09:41:12.9627264Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_367", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9628273Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9628537Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9628799Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9629482Z triton_mm_367 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9630532Z triton_mm_368 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9631593Z triton_mm_369 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9632744Z triton_mm_371 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9633792Z triton_mm_373 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9634843Z triton_mm_363 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9636093Z triton_mm_362 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9637211Z triton_mm_364 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9638683Z triton_mm_365 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9639916Z triton_mm_366 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9640827Z SingleProcess AUTOTUNE benchmarking takes 0.2034 seconds and 0.6302 seconds precompiling for 15 choices 2025-12-04T09:41:12.9641616Z Compiled module path: /tmp/tmpjh1d_rv8/lv/clvgfgovaqoxq6w2a3x3mdsjcbqczglzcgczjlxqwlgcbiom7udy.py 2025-12-04T09:41:12.9642412Z Compiled module path: /tmp/tmpjh1d_rv8/jj/cjjeeuywsjc5aexqz2oy52jb5r72fpvdjcb4gex7fwek3nrj3hsb.py 2025-12-04T09:41:12.9643127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9643509Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9643830Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9644316Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9646347Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9648287Z graph_break [] 2025-12-04T09:41:12.9648527Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9648915Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9649291Z Autotune Choices Stats: 2025-12-04T09:41:12.9650273Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_393", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9651285Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9651547Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9651808Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9652479Z triton_mm_393 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9653521Z triton_mm_396 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9654691Z triton_mm_403 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9655732Z triton_mm_395 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9656765Z triton_mm_401 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9657805Z triton_mm_391 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9658839Z triton_mm_392 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9659911Z triton_mm_394 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9660934Z triton_mm_397 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9661983Z triton_mm_398 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9662888Z SingleProcess AUTOTUNE benchmarking takes 0.2038 seconds and 0.6255 seconds precompiling for 15 choices 2025-12-04T09:41:12.9663694Z Compiled module path: /tmp/tmp23tycbp4/jm/cjmtzqck2innrgpl6fb7ywuwcpt4xhwrp3gneiorpkwkqojqod7x.py 2025-12-04T09:41:12.9664554Z Compiled module path: /tmp/tmp23tycbp4/jx/cjxhng63k4y7iiuaxtdkzcsiabmxi5ytaoafh4mubljd3jqi5aae.py 2025-12-04T09:41:12.9665202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9665581Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9665895Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9666373Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9668496Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9670443Z graph_break [] 2025-12-04T09:41:12.9670686Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9671060Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9671434Z Autotune Choices Stats: 2025-12-04T09:41:12.9672440Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_434", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9680466Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9680736Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9681008Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9681703Z triton_mm_434 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9682894Z triton_mm_427 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9683927Z triton_mm_431 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9684952Z triton_mm_423 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9685988Z triton_mm_421 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9687085Z triton_mm_424 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9688166Z triton_mm_426 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9689199Z triton_mm_430 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9690238Z triton_mm_432 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:12.9691277Z triton_mm_433 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9692241Z SingleProcess AUTOTUNE benchmarking takes 0.2047 seconds and 0.6105 seconds precompiling for 15 choices 2025-12-04T09:41:12.9693040Z Compiled module path: /tmp/tmphd5usaet/dc/cdcdxyzy6cd3jbgwclr23jsuvgkhqh5tbf32wst7ra4nnzfd3rty.py 2025-12-04T09:41:12.9693851Z Compiled module path: /tmp/tmphd5usaet/6k/c6kc65to4c26vezog4ygmrvhsa6qc6xsjj62rc72vinetzwzf6da.py 2025-12-04T09:41:12.9694587Z ______ BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code _______ 2025-12-04T09:41:12.9695071Z Traceback (most recent call last): 2025-12-04T09:41:12.9695671Z File "/var/lib/jenkins/workspace/test/inductor/test_benchmark_fusion.py", line 303, in test_equivalent_template_code 2025-12-04T09:41:12.9696364Z ).check("" if config.cpp_wrapper else "return").run(out_code[0]) 2025-12-04T09:41:12.9696894Z RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:12.9697338Z Searched string: 2025-12-04T09:41:12.9697587Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:12.9697899Z torch.cuda.set_device(0) 2025-12-04T09:41:12.9698272Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:12.9698787Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:12.9699237Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9699651Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:12.9700055Z del arg0_1 2025-12-04T09:41:12.9700283Z del arg2_1 2025-12-04T09:41:12.9700527Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:12.9700985Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:12.9701447Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9701846Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:12.9702234Z del arg1_1 2025-12-04T09:41:12.9702456Z return (buf1, ) 2025-12-04T09:41:12.9702617Z 2025-12-04T09:41:12.9702723Z runner = Runner(partitions=[]) 2025-12-04T09:41:12.9702984Z call = runner.call 2025-12-04T09:41:12.9703390Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:12.9703658Z 2025-12-04T09:41:12.9703663Z 2025-12-04T09:41:12.9703804Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:12.9704177Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:12.9704564Z from torch._inductor.utils import print_performance 2025-12-04T09:41:12.9705010Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9705517Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9706013Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9706433Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:12.9706817Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:12.9707130Z 2025-12-04T09:41:12.9707134Z 2025-12-04T09:41:12.9707227Z if __name__ == "__main__": 2025-12-04T09:41:12.9707602Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:12.9708065Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:12.9708447Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:12.9708662Z 2025-12-04T09:41:12.9708665Z 2025-12-04T09:41:12.9708855Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:12.9709788Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:12.9710540Z 2025-12-04T09:41:12.9710760Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:12.9711260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9711681Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9711993Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9713905Z inductor [('triton_bundler_save_kernel', 216), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('async_compile_cache_miss', 7), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9715916Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9716370Z graph_break [] 2025-12-04T09:41:12.9716611Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9716989Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9718596Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py:3433: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage() 2025-12-04T09:41:12.9720142Z current_size = base.storage().size() 2025-12-04T09:41:12.9720478Z Autotune Choices Stats: 2025-12-04T09:41:12.9721454Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_10", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9722472Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9722736Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9723011Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9723866Z triton_mm_10 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9724917Z triton_mm_9 0.0277 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9725948Z triton_mm_11 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9727025Z triton_mm_13 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9728066Z triton_mm_14 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9729139Z triton_mm_12 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:12.9730172Z triton_mm_1 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9731188Z triton_mm_2 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9732203Z triton_mm_6 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9733217Z triton_mm_7 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9734163Z SingleProcess AUTOTUNE benchmarking takes 0.1729 seconds and 1.0713 seconds precompiling for 15 choices 2025-12-04T09:41:12.9734971Z Compiled module path: /tmp/tmpgejy12ej/qv/cqvwpzgbltbzt3unpdzscv64a2elpufsmhey6btnq73pj2f4vaa7.py 2025-12-04T09:41:12.9736068Z Compiled module path: /tmp/tmpgejy12ej/jj/cjj4tsxlplxvzz5g4jfphbz5jsxl7b3fd5mxbhxre5lcbuh3vmrc.py 2025-12-04T09:41:12.9736700Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9737078Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9737387Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9737866Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9739891Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9741736Z graph_break [] 2025-12-04T09:41:12.9741969Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9742349Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9742716Z Autotune Choices Stats: 2025-12-04T09:41:12.9743681Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_30", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9744689Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9744946Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9745204Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9746014Z triton_mm_30 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9747051Z triton_mm_32 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9748073Z triton_mm_33 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9749088Z triton_mm_34 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9750159Z triton_mm_35 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9751249Z triton_mm_36 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9752276Z triton_mm_37 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9753305Z triton_mm_38 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9754339Z triton_mm_39 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9755437Z triton_mm_40 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9756344Z SingleProcess AUTOTUNE benchmarking takes 0.2013 seconds and 0.7407 seconds precompiling for 15 choices 2025-12-04T09:41:12.9757142Z Compiled module path: /tmp/tmpu9odmkxw/ai/cai5yet7zvfv3ry4tqny7aptct3cbabbgknxxt3otvygyekzraji.py 2025-12-04T09:41:12.9757949Z Compiled module path: /tmp/tmpu9odmkxw/nq/cnqm37sxiqmqocnfb7wdcdqxnd56ra26fhpzusixi763umecdh56.py 2025-12-04T09:41:12.9758569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9758947Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9759258Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9759736Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9761861Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 36), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 21), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 14), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9763793Z graph_break [] 2025-12-04T09:41:12.9764029Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9764406Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9764779Z Autotune Choices Stats: 2025-12-04T09:41:12.9765760Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_73", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9766770Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9767161Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9767434Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9768113Z triton_mm_73 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9769140Z triton_mm_62 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9770142Z triton_mm_65 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9771165Z triton_mm_71 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9772229Z triton_mm_61 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9773247Z triton_mm_63 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9774260Z triton_mm_60 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9775269Z triton_mm_64 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9776330Z triton_mm_67 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9777361Z triton_mm_68 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9778258Z SingleProcess AUTOTUNE benchmarking takes 0.2006 seconds and 0.6267 seconds precompiling for 15 choices 2025-12-04T09:41:12.9779046Z Compiled module path: /tmp/tmpqun5dtux/q7/cq7yaif2wknvncdje2hbdwcngroryoj2mtkk4dkwdkx45aat6hm7.py 2025-12-04T09:41:12.9779860Z Compiled module path: /tmp/tmpqun5dtux/y6/cy66mazx6rmemwkg6jecmgbfmpsvz6cfz5y45krjnl3azaph6y7t.py 2025-12-04T09:41:12.9780487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9780862Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9781168Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9781650Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9783688Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9785527Z graph_break [] 2025-12-04T09:41:12.9785764Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9786134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9786505Z Autotune Choices Stats: 2025-12-04T09:41:12.9787592Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_91", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9788608Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9788863Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9789125Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9789803Z triton_mm_91 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9790861Z triton_mm_95 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9791998Z triton_mm_90 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9793081Z triton_mm_92 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9794096Z triton_mm_93 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9795112Z triton_mm_96 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9796125Z triton_mm_97 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9797141Z triton_mm_98 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9798205Z triton_mm_94 0.0307 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9799287Z triton_mm_100 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9800198Z SingleProcess AUTOTUNE benchmarking takes 0.2153 seconds and 0.6426 seconds precompiling for 15 choices 2025-12-04T09:41:12.9800979Z Compiled module path: /tmp/tmp2v_kzp9a/fj/cfjz3b7ajfatqwomwryerntjnvk6kd5wysrcdign42gjpmvqkh3t.py 2025-12-04T09:41:12.9801773Z Compiled module path: /tmp/tmp2v_kzp9a/57/c57neznqb6ba2jtzlgbchzekwlmukri7wpo65rmsqusn7t6xlfxx.py 2025-12-04T09:41:12.9802392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9802771Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9803078Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9803560Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9805667Z inductor [('triton_bundler_save_kernel', 288), ('benchmarking.InductorBenchmarker.benchmark_gpu', 38), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 23), ('coordesc_tuning_bench', 16), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9807644Z graph_break [] 2025-12-04T09:41:12.9807874Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9808251Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9808624Z Autotune Choices Stats: 2025-12-04T09:41:12.9809683Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_134", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9810688Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9810950Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9811217Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9811889Z triton_mm_134 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9812941Z triton_mm_121 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9814023Z triton_mm_122 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9815044Z triton_mm_124 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9816065Z triton_mm_125 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9817080Z triton_mm_128 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9818115Z triton_mm_129 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9819194Z triton_mm_130 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9820222Z triton_mm_131 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9821236Z triton_mm_120 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9822138Z SingleProcess AUTOTUNE benchmarking takes 0.2007 seconds and 0.6630 seconds precompiling for 15 choices 2025-12-04T09:41:12.9822924Z Compiled module path: /tmp/tmp2w3f9rlp/jd/cjdbzzfn2s4jgc6gkasctk4suc4halinn3hncy6u26d22lzxdphd.py 2025-12-04T09:41:12.9823742Z Compiled module path: /tmp/tmp2w3f9rlp/4s/c4sivzup5qddnp7nobqsap2nr2ndjcdy3tashkjs4q3gubdqirss.py 2025-12-04T09:41:12.9824385Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9824756Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9825071Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9825548Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9827570Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9829404Z graph_break [] 2025-12-04T09:41:12.9829638Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9830014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9830465Z Autotune Choices Stats: 2025-12-04T09:41:12.9831444Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_158", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9832453Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9832710Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9832985Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9833648Z triton_mm_158 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9834686Z triton_mm_159 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9835987Z triton_mm_161 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9837017Z triton_mm_150 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9837479Z triton_mm_151 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9837940Z triton_mm_153 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9838482Z triton_mm_154 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9838944Z triton_mm_155 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9839403Z triton_mm_156 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9839864Z triton_mm_157 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9840202Z SingleProcess AUTOTUNE benchmarking takes 0.2054 seconds and 0.6100 seconds precompiling for 15 choices 2025-12-04T09:41:12.9840564Z Compiled module path: /tmp/tmpwi6t5aeb/gn/cgneadkz2tnedarmlzhqm4gn55whqeexaln2fzmkic7oxhyres4r.py 2025-12-04T09:41:12.9840915Z Compiled module path: /tmp/tmpwi6t5aeb/gy/cgyxpcavcgmmp26t6h56dechg7qqid3n7dyu5q6u5h3b6c7fepoj.py 2025-12-04T09:41:12.9841100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9841193Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9841331Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9841579Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9843261Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9843356Z graph_break [] 2025-12-04T09:41:12.9843458Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9843752Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9843844Z Autotune Choices Stats: 2025-12-04T09:41:12.9844664Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_186", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9844760Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9844844Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9844957Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9845425Z triton_mm_186 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9845958Z triton_mm_193 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9846431Z triton_mm_182 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9846904Z triton_mm_194 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9847443Z triton_mm_184 0.0308 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9847907Z triton_mm_187 0.0317 ms 90.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9848445Z triton_mm_185 0.0317 ms 90.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9848913Z triton_mm_180 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9849379Z triton_mm_181 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9849842Z triton_mm_183 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9850188Z SingleProcess AUTOTUNE benchmarking takes 0.2087 seconds and 0.6406 seconds precompiling for 15 choices 2025-12-04T09:41:12.9850552Z Compiled module path: /tmp/tmpk78l5vwh/kk/ckk35oaicdyefn5yev5gn7hgiqa5qquexctvfdcqebmw6q6whm6z.py 2025-12-04T09:41:12.9850904Z Compiled module path: /tmp/tmpk78l5vwh/hm/chmuaf5ovlicrcx6s7zj4pujz62mn7yxmmkafqcdbwqtfoym5ys5.py 2025-12-04T09:41:12.9851082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9851177Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9851309Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9851558Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9854200Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9854297Z graph_break [] 2025-12-04T09:41:12.9854406Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9854581Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9854677Z Autotune Choices Stats: 2025-12-04T09:41:12.9855503Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_220", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9855595Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9855682Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9855794Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9856326Z triton_mm_220 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9856800Z triton_mm_211 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9857263Z triton_mm_212 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9857729Z triton_mm_214 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9858192Z triton_mm_215 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9858710Z triton_mm_216 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9859179Z triton_mm_217 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9859644Z triton_mm_218 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9860115Z triton_mm_219 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9860581Z triton_mm_221 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9860929Z SingleProcess AUTOTUNE benchmarking takes 0.1998 seconds and 0.6553 seconds precompiling for 15 choices 2025-12-04T09:41:12.9861291Z Compiled module path: /tmp/tmpwuxhdg2k/ip/cipioi7ficmn3ww6fbrd2oyul2gc6oxnifncerak7mzlxvpbihdc.py 2025-12-04T09:41:12.9861656Z Compiled module path: /tmp/tmpwuxhdg2k/tq/ctqm7bgwweclnxyitnuulxziaydok5e5vcjg4prplje6cyryh3p3.py 2025-12-04T09:41:12.9861833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9861927Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9862064Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9862312Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9864079Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9864168Z graph_break [] 2025-12-04T09:41:12.9864272Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9864453Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9864544Z Autotune Choices Stats: 2025-12-04T09:41:12.9865372Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_250", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9865465Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9865587Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9865696Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9866174Z triton_mm_250 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9866635Z triton_mm_245 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9867104Z triton_mm_249 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9867571Z triton_mm_251 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9868080Z triton_mm_243 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9868551Z triton_mm_241 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9869016Z triton_mm_240 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9869474Z triton_mm_242 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9869953Z triton_mm_244 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9870450Z triton_mm_246 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9870790Z SingleProcess AUTOTUNE benchmarking takes 0.2009 seconds and 0.6388 seconds precompiling for 15 choices 2025-12-04T09:41:12.9871155Z Compiled module path: /tmp/tmpv4covkzm/xp/cxpol2uuscxtvrr3qpm7ebdm7l5updopjmnk75hpevamiwzj65o7.py 2025-12-04T09:41:12.9871502Z Compiled module path: /tmp/tmpv4covkzm/rn/crnctdx35xmvga3ti6xdrde3d3i7bt47oqtpexetsy437yfgpmyy.py 2025-12-04T09:41:12.9871683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9871774Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9871906Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9872159Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9873922Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9874011Z graph_break [] 2025-12-04T09:41:12.9874112Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9874284Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9874381Z Autotune Choices Stats: 2025-12-04T09:41:12.9875202Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_276", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.02768000029027462, "best_triton_pos": 0} 2025-12-04T09:41:12.9875342Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9875427Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9875539Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9876025Z triton_mm_276 0.0277 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9876490Z triton_mm_272 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9876953Z triton_mm_273 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9877419Z triton_mm_277 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9877937Z triton_mm_280 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9878405Z triton_mm_281 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9878868Z triton_mm_278 0.0296 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9879330Z triton_mm_270 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9879795Z triton_mm_271 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9880297Z triton_mm_274 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9880663Z SingleProcess AUTOTUNE benchmarking takes 0.2011 seconds and 0.6415 seconds precompiling for 15 choices 2025-12-04T09:41:12.9881012Z Compiled module path: /tmp/tmpuyrn9hru/66/c66x747yoejfwu2glewx7h3jab2qxz2k5kaaujahxklodqyj4cnu.py 2025-12-04T09:41:12.9881362Z Compiled module path: /tmp/tmpuyrn9hru/y7/cy7cdv3w3mgfm2t6nqc7yr2wtagkyvkt74hbgdknhwoqeeslovzq.py 2025-12-04T09:41:12.9881539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9881631Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9881768Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9882015Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9883873Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9883957Z graph_break [] 2025-12-04T09:41:12.9884070Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9884245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9884337Z Autotune Choices Stats: 2025-12-04T09:41:12.9885164Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_302", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9885323Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9885407Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9885514Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9885982Z triton_mm_302 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9886450Z triton_mm_303 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9886915Z triton_mm_305 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9887489Z triton_mm_307 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9887968Z triton_mm_309 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9888426Z triton_mm_300 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9888894Z triton_mm_301 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9889351Z triton_mm_304 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9889826Z triton_mm_306 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9890292Z triton_mm_308 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9890634Z SingleProcess AUTOTUNE benchmarking takes 0.2012 seconds and 0.6271 seconds precompiling for 15 choices 2025-12-04T09:41:12.9890994Z Compiled module path: /tmp/tmpnxfvglo0/mv/cmv5svnrqdgxfpijsujfb6lt3msltpw2rqji7ouqvhdeeykjqzmz.py 2025-12-04T09:41:12.9891343Z Compiled module path: /tmp/tmpnxfvglo0/rx/crxtbd5ejbufzmuyijmfl3i77mx7bc25cjvtl52cepcxot7477r3.py 2025-12-04T09:41:12.9891522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9891616Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9891750Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9892000Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9893809Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9893898Z graph_break [] 2025-12-04T09:41:12.9894002Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9894177Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9894280Z Autotune Choices Stats: 2025-12-04T09:41:12.9895228Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_333", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9895324Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9895410Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9895515Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9895992Z triton_mm_333 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9896461Z triton_mm_335 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9896991Z triton_mm_343 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9897466Z triton_mm_337 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9897936Z triton_mm_332 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9898405Z triton_mm_338 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9898870Z triton_mm_330 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9899344Z triton_mm_331 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9899817Z triton_mm_334 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9900277Z triton_mm_336 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9900614Z SingleProcess AUTOTUNE benchmarking takes 0.4010 seconds and 0.6248 seconds precompiling for 15 choices 2025-12-04T09:41:12.9900972Z Compiled module path: /tmp/tmp9c3250tk/oj/cojlvuknqajncjwhcs6yfkfjjtihtyenckj7m5l37blt326fq55c.py 2025-12-04T09:41:12.9901325Z Compiled module path: /tmp/tmp9c3250tk/st/cst5rwasvqebnwrew6gvackeisqs7qwte54od7766agjws6s3ura.py 2025-12-04T09:41:12.9901504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9901603Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9901819Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9902070Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9903755Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9903836Z graph_break [] 2025-12-04T09:41:12.9903948Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9904161Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9904252Z Autotune Choices Stats: 2025-12-04T09:41:12.9905087Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_367", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9905183Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9905274Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9905376Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9905846Z triton_mm_367 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9906320Z triton_mm_368 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9906897Z triton_mm_369 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9907415Z triton_mm_371 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9907893Z triton_mm_373 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9908398Z triton_mm_363 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9908855Z triton_mm_362 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9909322Z triton_mm_364 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9909789Z triton_mm_365 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9910252Z triton_mm_366 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9910593Z SingleProcess AUTOTUNE benchmarking takes 0.2034 seconds and 0.6302 seconds precompiling for 15 choices 2025-12-04T09:41:12.9910997Z Compiled module path: /tmp/tmpjh1d_rv8/lv/clvgfgovaqoxq6w2a3x3mdsjcbqczglzcgczjlxqwlgcbiom7udy.py 2025-12-04T09:41:12.9911340Z Compiled module path: /tmp/tmpjh1d_rv8/jj/cjjeeuywsjc5aexqz2oy52jb5r72fpvdjcb4gex7fwek3nrj3hsb.py 2025-12-04T09:41:12.9911601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9911696Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9911833Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9912079Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9913761Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9913889Z graph_break [] 2025-12-04T09:41:12.9913992Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9914177Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9914267Z Autotune Choices Stats: 2025-12-04T09:41:12.9915084Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_393", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9915182Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9915265Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9915380Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9915846Z triton_mm_393 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9916364Z triton_mm_396 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9916847Z triton_mm_403 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9917305Z triton_mm_395 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9917776Z triton_mm_401 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9918244Z triton_mm_391 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9918722Z triton_mm_392 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9919188Z triton_mm_394 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9919653Z triton_mm_397 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9920121Z triton_mm_398 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9920458Z SingleProcess AUTOTUNE benchmarking takes 0.2038 seconds and 0.6255 seconds precompiling for 15 choices 2025-12-04T09:41:12.9920826Z Compiled module path: /tmp/tmp23tycbp4/jm/cjmtzqck2innrgpl6fb7ywuwcpt4xhwrp3gneiorpkwkqojqod7x.py 2025-12-04T09:41:12.9921284Z Compiled module path: /tmp/tmp23tycbp4/jx/cjxhng63k4y7iiuaxtdkzcsiabmxi5ytaoafh4mubljd3jqi5aae.py 2025-12-04T09:41:12.9921461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9921557Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9921689Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9921938Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9923706Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9923849Z graph_break [] 2025-12-04T09:41:12.9923950Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9924124Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9924218Z Autotune Choices Stats: 2025-12-04T09:41:12.9925044Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_434", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9925135Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9925223Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9925327Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9925855Z triton_mm_434 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9926334Z triton_mm_427 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9926802Z triton_mm_431 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9927317Z triton_mm_423 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9927782Z triton_mm_421 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9928261Z triton_mm_424 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9928724Z triton_mm_426 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9929190Z triton_mm_430 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9929665Z triton_mm_432 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:12.9930181Z triton_mm_433 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9930534Z SingleProcess AUTOTUNE benchmarking takes 0.2047 seconds and 0.6105 seconds precompiling for 15 choices 2025-12-04T09:41:12.9930980Z Compiled module path: /tmp/tmphd5usaet/dc/cdcdxyzy6cd3jbgwclr23jsuvgkhqh5tbf32wst7ra4nnzfd3rty.py 2025-12-04T09:41:12.9931330Z Compiled module path: /tmp/tmphd5usaet/6k/c6kc65to4c26vezog4ygmrvhsa6qc6xsjj62rc72vinetzwzf6da.py 2025-12-04T09:41:12.9931504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9931597Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9931738Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9931987Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9933769Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9933893Z graph_break [] 2025-12-04T09:41:12.9933995Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9934180Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9934270Z Autotune Choices Stats: 2025-12-04T09:41:12.9935101Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_463", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9935630Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9935741Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9935854Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9936343Z triton_mm_463 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9936807Z triton_mm_450 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9937271Z triton_mm_451 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9937728Z triton_mm_453 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9938216Z triton_mm_464 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9938688Z triton_mm_455 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9939156Z triton_mm_458 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9939623Z triton_mm_460 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9940139Z triton_mm_452 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9940609Z triton_mm_454 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9941076Z SingleProcess AUTOTUNE benchmarking takes 0.2215 seconds and 0.6362 seconds precompiling for 15 choices 2025-12-04T09:41:12.9941443Z Compiled module path: /tmp/tmpc9pvlbau/hy/chya6s2u6d7btbkujfwyicdxobscvkdvifjkht7lz23hxz7lly2d.py 2025-12-04T09:41:12.9941802Z Compiled module path: /tmp/tmpc9pvlbau/jb/cjb6n2psnnfj2lf4fxigzmxslomoqidmvqxghrzqqz2wypsryyuy.py 2025-12-04T09:41:12.9942081Z ______ BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code _______ 2025-12-04T09:41:12.9942185Z Traceback (most recent call last): 2025-12-04T09:41:12.9942604Z File "/var/lib/jenkins/workspace/test/inductor/test_benchmark_fusion.py", line 303, in test_equivalent_template_code 2025-12-04T09:41:12.9942788Z ).check("" if config.cpp_wrapper else "return").run(out_code[0]) 2025-12-04T09:41:12.9943046Z RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:12.9943189Z Searched string: 2025-12-04T09:41:12.9943305Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:12.9943402Z torch.cuda.set_device(0) 2025-12-04T09:41:12.9943574Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:12.9943811Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:12.9943907Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9944129Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:12.9944211Z del arg0_1 2025-12-04T09:41:12.9944289Z del arg2_1 2025-12-04T09:41:12.9944391Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:12.9944655Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:12.9944813Z stream0 = get_raw_stream(0) 2025-12-04T09:41:12.9945017Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:12.9945102Z del arg1_1 2025-12-04T09:41:12.9945192Z return (buf1, ) 2025-12-04T09:41:12.9945197Z 2025-12-04T09:41:12.9945298Z runner = Runner(partitions=[]) 2025-12-04T09:41:12.9945378Z call = runner.call 2025-12-04T09:41:12.9945541Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:12.9945546Z 2025-12-04T09:41:12.9945550Z 2025-12-04T09:41:12.9945694Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:12.9945827Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:12.9945976Z from torch._inductor.utils import print_performance 2025-12-04T09:41:12.9946180Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9946380Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9946583Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:12.9946698Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:12.9946871Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:12.9946876Z 2025-12-04T09:41:12.9946880Z 2025-12-04T09:41:12.9946967Z if __name__ == "__main__": 2025-12-04T09:41:12.9947176Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:12.9947336Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:12.9947452Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:12.9947457Z 2025-12-04T09:41:12.9947461Z 2025-12-04T09:41:12.9947644Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:12.9948291Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:12.9948298Z 2025-12-04T09:41:12.9948527Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:12.9948785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9948879Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9949018Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9950699Z inductor [('triton_bundler_save_kernel', 216), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('async_compile_cache_miss', 7), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9950954Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9951074Z graph_break [] 2025-12-04T09:41:12.9951176Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9951363Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9952679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py:3433: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage() 2025-12-04T09:41:12.9952791Z current_size = base.storage().size() 2025-12-04T09:41:12.9952879Z Autotune Choices Stats: 2025-12-04T09:41:12.9953712Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_10", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9953846Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9953935Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9954043Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9954515Z triton_mm_10 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9954979Z triton_mm_9 0.0277 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9955445Z triton_mm_11 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9955913Z triton_mm_13 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9956390Z triton_mm_14 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9956853Z triton_mm_12 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:12.9961750Z triton_mm_1 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9962245Z triton_mm_2 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9962707Z triton_mm_6 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9963321Z triton_mm_7 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9963668Z SingleProcess AUTOTUNE benchmarking takes 0.1729 seconds and 1.0713 seconds precompiling for 15 choices 2025-12-04T09:41:12.9964040Z Compiled module path: /tmp/tmpgejy12ej/qv/cqvwpzgbltbzt3unpdzscv64a2elpufsmhey6btnq73pj2f4vaa7.py 2025-12-04T09:41:12.9964395Z Compiled module path: /tmp/tmpgejy12ej/jj/cjj4tsxlplxvzz5g4jfphbz5jsxl7b3fd5mxbhxre5lcbuh3vmrc.py 2025-12-04T09:41:12.9964574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9964675Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9964809Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9965070Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9966786Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9966871Z graph_break [] 2025-12-04T09:41:12.9967073Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9967253Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9967346Z Autotune Choices Stats: 2025-12-04T09:41:12.9968170Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_30", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9968341Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9968427Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9968536Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9969011Z triton_mm_30 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9969473Z triton_mm_32 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9969935Z triton_mm_33 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9970417Z triton_mm_34 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9970878Z triton_mm_35 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9971340Z triton_mm_36 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9971803Z triton_mm_37 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9972270Z triton_mm_38 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9972811Z triton_mm_39 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9973280Z triton_mm_40 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9973624Z SingleProcess AUTOTUNE benchmarking takes 0.2013 seconds and 0.7407 seconds precompiling for 15 choices 2025-12-04T09:41:12.9973984Z Compiled module path: /tmp/tmpu9odmkxw/ai/cai5yet7zvfv3ry4tqny7aptct3cbabbgknxxt3otvygyekzraji.py 2025-12-04T09:41:12.9974343Z Compiled module path: /tmp/tmpu9odmkxw/nq/cnqm37sxiqmqocnfb7wdcdqxnd56ra26fhpzusixi763umecdh56.py 2025-12-04T09:41:12.9974516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9974612Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9974788Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9975041Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9976812Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 36), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 21), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 14), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9976896Z graph_break [] 2025-12-04T09:41:12.9977004Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9977175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9977314Z Autotune Choices Stats: 2025-12-04T09:41:12.9978148Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_73", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9978241Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9978324Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9978440Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9978912Z triton_mm_73 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9979380Z triton_mm_62 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9979840Z triton_mm_65 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9980311Z triton_mm_71 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9980774Z triton_mm_61 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9981233Z triton_mm_63 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9981702Z triton_mm_60 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9982237Z triton_mm_64 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9982705Z triton_mm_67 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9983165Z triton_mm_68 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9983503Z SingleProcess AUTOTUNE benchmarking takes 0.2006 seconds and 0.6267 seconds precompiling for 15 choices 2025-12-04T09:41:12.9983868Z Compiled module path: /tmp/tmpqun5dtux/q7/cq7yaif2wknvncdje2hbdwcngroryoj2mtkk4dkwdkx45aat6hm7.py 2025-12-04T09:41:12.9984219Z Compiled module path: /tmp/tmpqun5dtux/y6/cy66mazx6rmemwkg6jecmgbfmpsvz6cfz5y45krjnl3azaph6y7t.py 2025-12-04T09:41:12.9984439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9984530Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9984667Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9984923Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9986608Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9986732Z graph_break [] 2025-12-04T09:41:12.9986835Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9987011Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9987107Z Autotune Choices Stats: 2025-12-04T09:41:12.9987931Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_91", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:12.9988026Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9988111Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9988213Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9988690Z triton_mm_91 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9989150Z triton_mm_95 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9989622Z triton_mm_90 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:12.9990076Z triton_mm_92 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9990550Z triton_mm_93 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9991001Z triton_mm_96 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:12.9991464Z triton_mm_97 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9992006Z triton_mm_98 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:12.9992463Z triton_mm_94 0.0307 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9992936Z triton_mm_100 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9993273Z SingleProcess AUTOTUNE benchmarking takes 0.2153 seconds and 0.6426 seconds precompiling for 15 choices 2025-12-04T09:41:12.9993625Z Compiled module path: /tmp/tmp2v_kzp9a/fj/cfjz3b7ajfatqwomwryerntjnvk6kd5wysrcdign42gjpmvqkh3t.py 2025-12-04T09:41:12.9994008Z Compiled module path: /tmp/tmp2v_kzp9a/57/c57neznqb6ba2jtzlgbchzekwlmukri7wpo65rmsqusn7t6xlfxx.py 2025-12-04T09:41:12.9994191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:12.9994285Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:12.9994416Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:12.9994668Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:12.9996430Z inductor [('triton_bundler_save_kernel', 288), ('benchmarking.InductorBenchmarker.benchmark_gpu', 38), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 23), ('coordesc_tuning_bench', 16), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:12.9996579Z graph_break [] 2025-12-04T09:41:12.9996683Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:12.9996861Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:12.9996955Z Autotune Choices Stats: 2025-12-04T09:41:12.9997786Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_134", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:12.9997877Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:12.9997969Z strides: [256, 1], [1, 256] 2025-12-04T09:41:12.9998074Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:12.9998555Z triton_mm_134 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:12.9999035Z triton_mm_121 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:12.9999505Z triton_mm_122 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:12.9999976Z triton_mm_124 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0000491Z triton_mm_125 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0000960Z triton_mm_128 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0001506Z triton_mm_129 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0001977Z triton_mm_130 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0002446Z triton_mm_131 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0002907Z triton_mm_120 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0003253Z SingleProcess AUTOTUNE benchmarking takes 0.2007 seconds and 0.6630 seconds precompiling for 15 choices 2025-12-04T09:41:13.0003649Z Compiled module path: /tmp/tmp2w3f9rlp/jd/cjdbzzfn2s4jgc6gkasctk4suc4halinn3hncy6u26d22lzxdphd.py 2025-12-04T09:41:13.0004002Z Compiled module path: /tmp/tmp2w3f9rlp/4s/c4sivzup5qddnp7nobqsap2nr2ndjcdy3tashkjs4q3gubdqirss.py 2025-12-04T09:41:13.0004182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0004277Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0004412Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0004658Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0006343Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0006473Z graph_break [] 2025-12-04T09:41:13.0006576Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0006753Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0006844Z Autotune Choices Stats: 2025-12-04T09:41:13.0007712Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_158", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0007814Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0007898Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0008011Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0008483Z triton_mm_158 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0008967Z triton_mm_159 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0009450Z triton_mm_161 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0009913Z triton_mm_150 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0010386Z triton_mm_151 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0010935Z triton_mm_153 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0011402Z triton_mm_154 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0011858Z triton_mm_155 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0012314Z triton_mm_156 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0012796Z triton_mm_157 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0013197Z SingleProcess AUTOTUNE benchmarking takes 0.2054 seconds and 0.6100 seconds precompiling for 15 choices 2025-12-04T09:41:13.0013564Z Compiled module path: /tmp/tmpwi6t5aeb/gn/cgneadkz2tnedarmlzhqm4gn55whqeexaln2fzmkic7oxhyres4r.py 2025-12-04T09:41:13.0013977Z Compiled module path: /tmp/tmpwi6t5aeb/gy/cgyxpcavcgmmp26t6h56dechg7qqid3n7dyu5q6u5h3b6c7fepoj.py 2025-12-04T09:41:13.0014225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0014357Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0014495Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0014748Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0016501Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0016656Z graph_break [] 2025-12-04T09:41:13.0016768Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0016942Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0017035Z Autotune Choices Stats: 2025-12-04T09:41:13.0017852Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_186", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0017948Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0018038Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0018149Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0018633Z triton_mm_186 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0019104Z triton_mm_193 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0019567Z triton_mm_182 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0020036Z triton_mm_194 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0020496Z triton_mm_184 0.0308 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0021046Z triton_mm_187 0.0317 ms 90.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0021508Z triton_mm_185 0.0317 ms 90.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0021970Z triton_mm_180 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0022434Z triton_mm_181 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0022895Z triton_mm_183 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0023286Z SingleProcess AUTOTUNE benchmarking takes 0.2087 seconds and 0.6406 seconds precompiling for 15 choices 2025-12-04T09:41:13.0023641Z Compiled module path: /tmp/tmpk78l5vwh/kk/ckk35oaicdyefn5yev5gn7hgiqa5qquexctvfdcqebmw6q6whm6z.py 2025-12-04T09:41:13.0023994Z Compiled module path: /tmp/tmpk78l5vwh/hm/chmuaf5ovlicrcx6s7zj4pujz62mn7yxmmkafqcdbwqtfoym5ys5.py 2025-12-04T09:41:13.0024168Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0024261Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0024400Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0024646Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0026374Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0026460Z graph_break [] 2025-12-04T09:41:13.0026562Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0026741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0026830Z Autotune Choices Stats: 2025-12-04T09:41:13.0027658Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_220", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0027753Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0027839Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0027951Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0028425Z triton_mm_220 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0028897Z triton_mm_211 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0029353Z triton_mm_212 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0029816Z triton_mm_214 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0030362Z triton_mm_215 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0030827Z triton_mm_216 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0031292Z triton_mm_217 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0031755Z triton_mm_218 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0032219Z triton_mm_219 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0032754Z triton_mm_221 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0033096Z SingleProcess AUTOTUNE benchmarking takes 0.1998 seconds and 0.6553 seconds precompiling for 15 choices 2025-12-04T09:41:13.0033460Z Compiled module path: /tmp/tmpwuxhdg2k/ip/cipioi7ficmn3ww6fbrd2oyul2gc6oxnifncerak7mzlxvpbihdc.py 2025-12-04T09:41:13.0033821Z Compiled module path: /tmp/tmpwuxhdg2k/tq/ctqm7bgwweclnxyitnuulxziaydok5e5vcjg4prplje6cyryh3p3.py 2025-12-04T09:41:13.0033998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0034090Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0034222Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0034517Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0036652Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0036749Z graph_break [] 2025-12-04T09:41:13.0036853Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0037030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0037125Z Autotune Choices Stats: 2025-12-04T09:41:13.0037955Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_250", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0038054Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0038139Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0038245Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0038730Z triton_mm_250 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0039189Z triton_mm_245 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0039661Z triton_mm_249 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0040264Z triton_mm_251 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0040723Z triton_mm_243 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0041194Z triton_mm_241 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0041647Z triton_mm_240 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0042111Z triton_mm_242 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0042632Z triton_mm_244 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0043094Z triton_mm_246 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0043435Z SingleProcess AUTOTUNE benchmarking takes 0.2009 seconds and 0.6388 seconds precompiling for 15 choices 2025-12-04T09:41:13.0043802Z Compiled module path: /tmp/tmpv4covkzm/xp/cxpol2uuscxtvrr3qpm7ebdm7l5updopjmnk75hpevamiwzj65o7.py 2025-12-04T09:41:13.0044158Z Compiled module path: /tmp/tmpv4covkzm/rn/crnctdx35xmvga3ti6xdrde3d3i7bt47oqtpexetsy437yfgpmyy.py 2025-12-04T09:41:13.0044330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0044485Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0044618Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0044868Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0046551Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0046634Z graph_break [] 2025-12-04T09:41:13.0046743Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0046918Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0047059Z Autotune Choices Stats: 2025-12-04T09:41:13.0047893Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_276", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.02768000029027462, "best_triton_pos": 0} 2025-12-04T09:41:13.0047985Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0048072Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0048182Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0048655Z triton_mm_276 0.0277 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0049129Z triton_mm_272 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0049597Z triton_mm_273 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0050199Z triton_mm_277 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0050675Z triton_mm_280 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0051143Z triton_mm_281 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0051609Z triton_mm_278 0.0296 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0052073Z triton_mm_270 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0052585Z triton_mm_271 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0053047Z triton_mm_274 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0053386Z SingleProcess AUTOTUNE benchmarking takes 0.2011 seconds and 0.6415 seconds precompiling for 15 choices 2025-12-04T09:41:13.0053736Z Compiled module path: /tmp/tmpuyrn9hru/66/c66x747yoejfwu2glewx7h3jab2qxz2k5kaaujahxklodqyj4cnu.py 2025-12-04T09:41:13.0054084Z Compiled module path: /tmp/tmpuyrn9hru/y7/cy7cdv3w3mgfm2t6nqc7yr2wtagkyvkt74hbgdknhwoqeeslovzq.py 2025-12-04T09:41:13.0054300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0054395Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0054534Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0054794Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0056470Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0056566Z graph_break [] 2025-12-04T09:41:13.0056672Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0056851Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0056943Z Autotune Choices Stats: 2025-12-04T09:41:13.0057768Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_302", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0057863Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0057947Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0058054Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0058527Z triton_mm_302 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0058992Z triton_mm_303 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0059551Z triton_mm_305 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0060039Z triton_mm_307 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0060549Z triton_mm_309 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0061009Z triton_mm_300 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0061476Z triton_mm_301 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0061991Z triton_mm_304 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0062453Z triton_mm_306 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0062918Z triton_mm_308 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0063258Z SingleProcess AUTOTUNE benchmarking takes 0.2012 seconds and 0.6271 seconds precompiling for 15 choices 2025-12-04T09:41:13.0063625Z Compiled module path: /tmp/tmpnxfvglo0/mv/cmv5svnrqdgxfpijsujfb6lt3msltpw2rqji7ouqvhdeeykjqzmz.py 2025-12-04T09:41:13.0064018Z Compiled module path: /tmp/tmpnxfvglo0/rx/crxtbd5ejbufzmuyijmfl3i77mx7bc25cjvtl52cepcxot7477r3.py 2025-12-04T09:41:13.0064193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0064295Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0064427Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0064676Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0066361Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0066455Z graph_break [] 2025-12-04T09:41:13.0066560Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0066745Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0066836Z Autotune Choices Stats: 2025-12-04T09:41:13.0067661Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_333", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0067756Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0067850Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0067957Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0068422Z triton_mm_333 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0068906Z triton_mm_335 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0069485Z triton_mm_343 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0069959Z triton_mm_337 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0070429Z triton_mm_332 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0070902Z triton_mm_338 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0071410Z triton_mm_330 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0071873Z triton_mm_331 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0072344Z triton_mm_334 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0072801Z triton_mm_336 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0073141Z SingleProcess AUTOTUNE benchmarking takes 0.4010 seconds and 0.6248 seconds precompiling for 15 choices 2025-12-04T09:41:13.0073534Z Compiled module path: /tmp/tmp9c3250tk/oj/cojlvuknqajncjwhcs6yfkfjjtihtyenckj7m5l37blt326fq55c.py 2025-12-04T09:41:13.0073903Z Compiled module path: /tmp/tmp9c3250tk/st/cst5rwasvqebnwrew6gvackeisqs7qwte54od7766agjws6s3ura.py 2025-12-04T09:41:13.0074077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0074169Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0074308Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0074560Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0076230Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0076317Z graph_break [] 2025-12-04T09:41:13.0076427Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0076608Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0076700Z Autotune Choices Stats: 2025-12-04T09:41:13.0077523Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_367", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0077623Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0077707Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0077820Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0078298Z triton_mm_367 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0078867Z triton_mm_368 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0079344Z triton_mm_369 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0079814Z triton_mm_371 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0080347Z triton_mm_373 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0080807Z triton_mm_363 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0081316Z triton_mm_362 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0081772Z triton_mm_364 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0082231Z triton_mm_365 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0082698Z triton_mm_366 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0083072Z SingleProcess AUTOTUNE benchmarking takes 0.2034 seconds and 0.6302 seconds precompiling for 15 choices 2025-12-04T09:41:13.0083440Z Compiled module path: /tmp/tmpjh1d_rv8/lv/clvgfgovaqoxq6w2a3x3mdsjcbqczglzcgczjlxqwlgcbiom7udy.py 2025-12-04T09:41:13.0083779Z Compiled module path: /tmp/tmpjh1d_rv8/jj/cjjeeuywsjc5aexqz2oy52jb5r72fpvdjcb4gex7fwek3nrj3hsb.py 2025-12-04T09:41:13.0083952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0084051Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0084184Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0084438Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0086108Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0086203Z graph_break [] 2025-12-04T09:41:13.0086312Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0086485Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0086577Z Autotune Choices Stats: 2025-12-04T09:41:13.0087440Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_393", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0087532Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0087629Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0087736Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0088293Z triton_mm_393 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0088760Z triton_mm_396 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0089242Z triton_mm_403 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0089712Z triton_mm_395 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0090210Z triton_mm_401 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0090818Z triton_mm_391 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0091280Z triton_mm_392 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0091739Z triton_mm_394 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0092201Z triton_mm_397 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0092708Z triton_mm_398 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0093062Z SingleProcess AUTOTUNE benchmarking takes 0.2038 seconds and 0.6255 seconds precompiling for 15 choices 2025-12-04T09:41:13.0093424Z Compiled module path: /tmp/tmp23tycbp4/jm/cjmtzqck2innrgpl6fb7ywuwcpt4xhwrp3gneiorpkwkqojqod7x.py 2025-12-04T09:41:13.0093778Z Compiled module path: /tmp/tmp23tycbp4/jx/cjxhng63k4y7iiuaxtdkzcsiabmxi5ytaoafh4mubljd3jqi5aae.py 2025-12-04T09:41:13.0093950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0094044Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0094183Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0094429Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0096198Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0096284Z graph_break [] 2025-12-04T09:41:13.0096389Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0096570Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0096660Z Autotune Choices Stats: 2025-12-04T09:41:13.0097490Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_434", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0097584Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0097666Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0097853Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0098332Z triton_mm_434 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0098801Z triton_mm_427 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0099266Z triton_mm_431 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0099722Z triton_mm_423 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0100251Z triton_mm_421 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0100712Z triton_mm_424 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0101170Z triton_mm_426 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0101633Z triton_mm_430 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0102110Z triton_mm_432 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0102625Z triton_mm_433 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0102967Z SingleProcess AUTOTUNE benchmarking takes 0.2047 seconds and 0.6105 seconds precompiling for 15 choices 2025-12-04T09:41:13.0103326Z Compiled module path: /tmp/tmphd5usaet/dc/cdcdxyzy6cd3jbgwclr23jsuvgkhqh5tbf32wst7ra4nnzfd3rty.py 2025-12-04T09:41:13.0103672Z Compiled module path: /tmp/tmphd5usaet/6k/c6kc65to4c26vezog4ygmrvhsa6qc6xsjj62rc72vinetzwzf6da.py 2025-12-04T09:41:13.0103852Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0103944Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0104077Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0104329Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0106098Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0106184Z graph_break [] 2025-12-04T09:41:13.0106289Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0106467Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0106558Z Autotune Choices Stats: 2025-12-04T09:41:13.0107479Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_463", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0107581Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0107665Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0107770Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0108256Z triton_mm_463 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0108718Z triton_mm_450 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0109186Z triton_mm_451 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0109691Z triton_mm_453 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0110211Z triton_mm_464 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0110666Z triton_mm_455 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0111131Z triton_mm_458 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0111603Z triton_mm_460 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0112108Z triton_mm_452 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0112576Z triton_mm_454 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0112916Z SingleProcess AUTOTUNE benchmarking takes 0.2215 seconds and 0.6362 seconds precompiling for 15 choices 2025-12-04T09:41:13.0113280Z Compiled module path: /tmp/tmpc9pvlbau/hy/chya6s2u6d7btbkujfwyicdxobscvkdvifjkht7lz23hxz7lly2d.py 2025-12-04T09:41:13.0113643Z Compiled module path: /tmp/tmpc9pvlbau/jb/cjb6n2psnnfj2lf4fxigzmxslomoqidmvqxghrzqqz2wypsryyuy.py 2025-12-04T09:41:13.0113818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0113923Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0114056Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0114309Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0115983Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0116064Z graph_break [] 2025-12-04T09:41:13.0116169Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0116342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0116434Z Autotune Choices Stats: 2025-12-04T09:41:13.0117336Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_484", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0117429Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0117516Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0117621Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0118091Z triton_mm_484 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0118564Z triton_mm_485 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0119030Z triton_mm_486 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0119548Z triton_mm_487 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0120030Z triton_mm_488 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0120538Z triton_mm_489 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0121004Z triton_mm_481 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0121503Z triton_mm_482 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0121972Z triton_mm_483 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0122428Z triton_mm_480 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0122766Z SingleProcess AUTOTUNE benchmarking takes 0.2056 seconds and 0.6477 seconds precompiling for 15 choices 2025-12-04T09:41:13.0123127Z Compiled module path: /tmp/tmp7po455mk/zq/czqdrwg6wlfanobk5rsuxzsuswzxqmizjrabbyvkkjoahvmngyue.py 2025-12-04T09:41:13.0123484Z Compiled module path: /tmp/tmp7po455mk/hc/chchsjouzzt5jvril5awri5ugmn3hzvu3i7mr4opswkq5aqbbj7y.py 2025-12-04T09:41:13.0123663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0123757Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0123900Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0124145Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0125818Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0125905Z graph_break [] 2025-12-04T09:41:13.0126010Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0126194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0126283Z Autotune Choices Stats: 2025-12-04T09:41:13.0127226Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_521", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0127324Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0127407Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0127514Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0127989Z triton_mm_521 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0128460Z triton_mm_524 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0128976Z triton_mm_517 0.0277 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0129439Z triton_mm_515 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0129905Z triton_mm_520 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0130375Z triton_mm_523 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0130885Z triton_mm_511 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0131351Z triton_mm_512 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0131809Z triton_mm_513 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0132272Z triton_mm_518 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0132608Z SingleProcess AUTOTUNE benchmarking takes 0.2023 seconds and 0.6312 seconds precompiling for 15 choices 2025-12-04T09:41:13.0132956Z Compiled module path: /tmp/tmpe_p5bjio/6o/c6ojtkxpkeubqosthw3efvd63vixnbse24c2gehuwspqfp2fx4at.py 2025-12-04T09:41:13.0133293Z Compiled module path: /tmp/tmpe_p5bjio/ct/cct6e5xbjmvmf7fbjj3pm3amu4sajwis4x3cmls64nm7u3zk74l7.py 2025-12-04T09:41:13.0133473Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0133573Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0133705Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0133955Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0136083Z inductor [('triton_bundler_save_kernel', 304), ('benchmarking.InductorBenchmarker.benchmark_gpu', 39), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 24), ('coordesc_tuning_bench', 17), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0136184Z graph_break [] 2025-12-04T09:41:13.0136286Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0136586Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0136683Z Autotune Choices Stats: 2025-12-04T09:41:13.0137508Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_553", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0137603Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0137697Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0137802Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0138282Z triton_mm_553 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0138819Z triton_mm_554 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0139285Z triton_mm_547 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0139757Z triton_mm_548 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0140222Z triton_mm_551 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0140684Z triton_mm_542 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0141232Z triton_mm_541 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0141699Z triton_mm_545 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0142156Z triton_mm_540 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0142612Z triton_mm_543 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0142954Z SingleProcess AUTOTUNE benchmarking takes 0.2014 seconds and 0.6266 seconds precompiling for 15 choices 2025-12-04T09:41:13.0143311Z Compiled module path: /tmp/tmpg8i5cgkd/em/cem6yffkh3my5x7um7rd7ufbeqaw7t6yzo4e75mwl2hys3ne5xbk.py 2025-12-04T09:41:13.0143671Z Compiled module path: /tmp/tmpg8i5cgkd/fy/cfyheiorobaq6ecqx7rnimkff2o3p2venjydoamv4ukkzofqfnc6.py 2025-12-04T09:41:13.0143944Z ______ BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code _______ 2025-12-04T09:41:13.0144050Z Traceback (most recent call last): 2025-12-04T09:41:13.0144470Z File "/var/lib/jenkins/workspace/test/inductor/test_benchmark_fusion.py", line 303, in test_equivalent_template_code 2025-12-04T09:41:13.0144645Z ).check("" if config.cpp_wrapper else "return").run(out_code[0]) 2025-12-04T09:41:13.0144904Z RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0144985Z Searched string: 2025-12-04T09:41:13.0145093Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0145196Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0145366Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0145610Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0145796Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0146008Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0146092Z del arg0_1 2025-12-04T09:41:13.0146172Z del arg2_1 2025-12-04T09:41:13.0146271Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0146535Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0146631Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0146828Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0146910Z del arg1_1 2025-12-04T09:41:13.0146991Z return (buf1, ) 2025-12-04T09:41:13.0147000Z 2025-12-04T09:41:13.0147099Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0147231Z call = runner.call 2025-12-04T09:41:13.0147393Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0147404Z 2025-12-04T09:41:13.0147408Z 2025-12-04T09:41:13.0147555Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0147686Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0147836Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0148046Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0148235Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0148433Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0148554Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0148720Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0148767Z 2025-12-04T09:41:13.0148773Z 2025-12-04T09:41:13.0148866Z if __name__ == "__main__": 2025-12-04T09:41:13.0149072Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0149239Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0149359Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0149363Z 2025-12-04T09:41:13.0149367Z 2025-12-04T09:41:13.0149544Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0150248Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0150253Z 2025-12-04T09:41:13.0150472Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0150648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0150749Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0150886Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0152572Z inductor [('triton_bundler_save_kernel', 216), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('async_compile_cache_miss', 7), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0152820Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0152903Z graph_break [] 2025-12-04T09:41:13.0153005Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0153177Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0154577Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py:3433: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage() 2025-12-04T09:41:13.0154693Z current_size = base.storage().size() 2025-12-04T09:41:13.0154785Z Autotune Choices Stats: 2025-12-04T09:41:13.0155608Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_10", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0155699Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0155790Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0155902Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0156374Z triton_mm_10 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0156889Z triton_mm_9 0.0277 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0157351Z triton_mm_11 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0157819Z triton_mm_13 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0158284Z triton_mm_14 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0158803Z triton_mm_12 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0159267Z triton_mm_1 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0159721Z triton_mm_2 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0160174Z triton_mm_6 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0160682Z triton_mm_7 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0161028Z SingleProcess AUTOTUNE benchmarking takes 0.1729 seconds and 1.0713 seconds precompiling for 15 choices 2025-12-04T09:41:13.0161392Z Compiled module path: /tmp/tmpgejy12ej/qv/cqvwpzgbltbzt3unpdzscv64a2elpufsmhey6btnq73pj2f4vaa7.py 2025-12-04T09:41:13.0161750Z Compiled module path: /tmp/tmpgejy12ej/jj/cjj4tsxlplxvzz5g4jfphbz5jsxl7b3fd5mxbhxre5lcbuh3vmrc.py 2025-12-04T09:41:13.0161925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0162017Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0162152Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0162398Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0164167Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0164255Z graph_break [] 2025-12-04T09:41:13.0164357Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0164533Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0164625Z Autotune Choices Stats: 2025-12-04T09:41:13.0165450Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_30", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0165549Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0165633Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0165796Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0166266Z triton_mm_30 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0166731Z triton_mm_32 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0167248Z triton_mm_33 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0167710Z triton_mm_34 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0168224Z triton_mm_35 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0168688Z triton_mm_36 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0169154Z triton_mm_37 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0169621Z triton_mm_38 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0170095Z triton_mm_39 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0170567Z triton_mm_40 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0170914Z SingleProcess AUTOTUNE benchmarking takes 0.2013 seconds and 0.7407 seconds precompiling for 15 choices 2025-12-04T09:41:13.0171277Z Compiled module path: /tmp/tmpu9odmkxw/ai/cai5yet7zvfv3ry4tqny7aptct3cbabbgknxxt3otvygyekzraji.py 2025-12-04T09:41:13.0171629Z Compiled module path: /tmp/tmpu9odmkxw/nq/cnqm37sxiqmqocnfb7wdcdqxnd56ra26fhpzusixi763umecdh56.py 2025-12-04T09:41:13.0171805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0171897Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0172027Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0172275Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0174115Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 36), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 21), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 14), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0174203Z graph_break [] 2025-12-04T09:41:13.0174306Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0174479Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0174571Z Autotune Choices Stats: 2025-12-04T09:41:13.0175393Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_73", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0175558Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0175648Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0175754Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0176232Z triton_mm_73 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0176687Z triton_mm_62 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0177146Z triton_mm_65 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0177614Z triton_mm_71 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0178120Z triton_mm_61 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0178582Z triton_mm_63 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0179036Z triton_mm_60 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0179495Z triton_mm_64 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0179957Z triton_mm_67 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0180424Z triton_mm_68 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0180758Z SingleProcess AUTOTUNE benchmarking takes 0.2006 seconds and 0.6267 seconds precompiling for 15 choices 2025-12-04T09:41:13.0181114Z Compiled module path: /tmp/tmpqun5dtux/q7/cq7yaif2wknvncdje2hbdwcngroryoj2mtkk4dkwdkx45aat6hm7.py 2025-12-04T09:41:13.0181471Z Compiled module path: /tmp/tmpqun5dtux/y6/cy66mazx6rmemwkg6jecmgbfmpsvz6cfz5y45krjnl3azaph6y7t.py 2025-12-04T09:41:13.0181647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0181738Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0181876Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0182124Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0183878Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0183962Z graph_break [] 2025-12-04T09:41:13.0184067Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0184240Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0184331Z Autotune Choices Stats: 2025-12-04T09:41:13.0185157Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_91", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0185290Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0185373Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0185483Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0185950Z triton_mm_91 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0186412Z triton_mm_95 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0186865Z triton_mm_90 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0187370Z triton_mm_92 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0187825Z triton_mm_93 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0188282Z triton_mm_96 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0188745Z triton_mm_97 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0189204Z triton_mm_98 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0189672Z triton_mm_94 0.0307 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0190172Z triton_mm_100 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0190525Z SingleProcess AUTOTUNE benchmarking takes 0.2153 seconds and 0.6426 seconds precompiling for 15 choices 2025-12-04T09:41:13.0190877Z Compiled module path: /tmp/tmp2v_kzp9a/fj/cfjz3b7ajfatqwomwryerntjnvk6kd5wysrcdign42gjpmvqkh3t.py 2025-12-04T09:41:13.0191216Z Compiled module path: /tmp/tmp2v_kzp9a/57/c57neznqb6ba2jtzlgbchzekwlmukri7wpo65rmsqusn7t6xlfxx.py 2025-12-04T09:41:13.0191394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0191489Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0191625Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0191875Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0193708Z inductor [('triton_bundler_save_kernel', 288), ('benchmarking.InductorBenchmarker.benchmark_gpu', 38), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 23), ('coordesc_tuning_bench', 16), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0193798Z graph_break [] 2025-12-04T09:41:13.0193901Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0194075Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0194212Z Autotune Choices Stats: 2025-12-04T09:41:13.0195052Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_134", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0195155Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0195239Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0195346Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0195830Z triton_mm_134 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0196295Z triton_mm_121 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0196799Z triton_mm_122 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0197271Z triton_mm_124 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0197742Z triton_mm_125 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0198209Z triton_mm_128 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0198673Z triton_mm_129 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0199161Z triton_mm_130 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0199625Z triton_mm_131 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0200117Z triton_mm_120 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0200481Z SingleProcess AUTOTUNE benchmarking takes 0.2007 seconds and 0.6630 seconds precompiling for 15 choices 2025-12-04T09:41:13.0200837Z Compiled module path: /tmp/tmp2w3f9rlp/jd/cjdbzzfn2s4jgc6gkasctk4suc4halinn3hncy6u26d22lzxdphd.py 2025-12-04T09:41:13.0201199Z Compiled module path: /tmp/tmp2w3f9rlp/4s/c4sivzup5qddnp7nobqsap2nr2ndjcdy3tashkjs4q3gubdqirss.py 2025-12-04T09:41:13.0201377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0201477Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0201685Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0201934Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0203608Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0203692Z graph_break [] 2025-12-04T09:41:13.0203797Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0204011Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0204106Z Autotune Choices Stats: 2025-12-04T09:41:13.0204938Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_158", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0205033Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0205120Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0205225Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0205695Z triton_mm_158 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0206177Z triton_mm_159 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0206703Z triton_mm_161 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0207215Z triton_mm_150 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0207684Z triton_mm_151 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0208151Z triton_mm_153 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0208611Z triton_mm_154 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0209077Z triton_mm_155 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0209535Z triton_mm_156 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0209996Z triton_mm_157 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0210337Z SingleProcess AUTOTUNE benchmarking takes 0.2054 seconds and 0.6100 seconds precompiling for 15 choices 2025-12-04T09:41:13.0210702Z Compiled module path: /tmp/tmpwi6t5aeb/gn/cgneadkz2tnedarmlzhqm4gn55whqeexaln2fzmkic7oxhyres4r.py 2025-12-04T09:41:13.0211052Z Compiled module path: /tmp/tmpwi6t5aeb/gy/cgyxpcavcgmmp26t6h56dechg7qqid3n7dyu5q6u5h3b6c7fepoj.py 2025-12-04T09:41:13.0211332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0211428Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0211566Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0211814Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0213488Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0213616Z graph_break [] 2025-12-04T09:41:13.0213719Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0213907Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0214001Z Autotune Choices Stats: 2025-12-04T09:41:13.0214813Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_186", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0214911Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0214995Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0215103Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0215569Z triton_mm_186 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0216088Z triton_mm_193 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0216552Z triton_mm_182 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0217020Z triton_mm_194 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0217479Z triton_mm_184 0.0308 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0217941Z triton_mm_187 0.0317 ms 90.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0218410Z triton_mm_185 0.0317 ms 90.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0218865Z triton_mm_180 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0219327Z triton_mm_181 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0219800Z triton_mm_183 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0220138Z SingleProcess AUTOTUNE benchmarking takes 0.2087 seconds and 0.6406 seconds precompiling for 15 choices 2025-12-04T09:41:13.0220508Z Compiled module path: /tmp/tmpk78l5vwh/kk/ckk35oaicdyefn5yev5gn7hgiqa5qquexctvfdcqebmw6q6whm6z.py 2025-12-04T09:41:13.0220932Z Compiled module path: /tmp/tmpk78l5vwh/hm/chmuaf5ovlicrcx6s7zj4pujz62mn7yxmmkafqcdbwqtfoym5ys5.py 2025-12-04T09:41:13.0221107Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0221204Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0221335Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0221586Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0223254Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0223385Z graph_break [] 2025-12-04T09:41:13.0223488Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0223661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0223754Z Autotune Choices Stats: 2025-12-04T09:41:13.0224573Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_220", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0224664Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0224751Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0224856Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0225375Z triton_mm_220 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0225847Z triton_mm_211 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0226309Z triton_mm_212 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0226773Z triton_mm_214 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0227231Z triton_mm_215 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0227693Z triton_mm_216 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0228163Z triton_mm_217 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0228630Z triton_mm_218 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0229092Z triton_mm_219 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0229554Z triton_mm_221 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0229899Z SingleProcess AUTOTUNE benchmarking takes 0.1998 seconds and 0.6553 seconds precompiling for 15 choices 2025-12-04T09:41:13.0230331Z Compiled module path: /tmp/tmpwuxhdg2k/ip/cipioi7ficmn3ww6fbrd2oyul2gc6oxnifncerak7mzlxvpbihdc.py 2025-12-04T09:41:13.0230693Z Compiled module path: /tmp/tmpwuxhdg2k/tq/ctqm7bgwweclnxyitnuulxziaydok5e5vcjg4prplje6cyryh3p3.py 2025-12-04T09:41:13.0230865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0230959Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0231098Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0242302Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0244531Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0244755Z graph_break [] 2025-12-04T09:41:13.0244871Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0245064Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0245169Z Autotune Choices Stats: 2025-12-04T09:41:13.0246018Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_250", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0246178Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0246267Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0246398Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0246887Z triton_mm_250 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0247486Z triton_mm_245 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0247954Z triton_mm_249 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0248447Z triton_mm_251 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0248911Z triton_mm_243 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0249383Z triton_mm_241 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0249864Z triton_mm_240 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0250357Z triton_mm_242 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0250830Z triton_mm_244 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0251292Z triton_mm_246 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0251824Z SingleProcess AUTOTUNE benchmarking takes 0.2009 seconds and 0.6388 seconds precompiling for 15 choices 2025-12-04T09:41:13.0252203Z Compiled module path: /tmp/tmpv4covkzm/xp/cxpol2uuscxtvrr3qpm7ebdm7l5updopjmnk75hpevamiwzj65o7.py 2025-12-04T09:41:13.0252556Z Compiled module path: /tmp/tmpv4covkzm/rn/crnctdx35xmvga3ti6xdrde3d3i7bt47oqtpexetsy437yfgpmyy.py 2025-12-04T09:41:13.0252740Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0252839Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0252975Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0253233Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0254910Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0255047Z graph_break [] 2025-12-04T09:41:13.0255156Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0255334Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0255441Z Autotune Choices Stats: 2025-12-04T09:41:13.0256263Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_276", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.02768000029027462, "best_triton_pos": 0} 2025-12-04T09:41:13.0256410Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0256495Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0256614Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0257095Z triton_mm_276 0.0277 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0257560Z triton_mm_272 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0258033Z triton_mm_273 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0258503Z triton_mm_277 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0259049Z triton_mm_280 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0259658Z triton_mm_281 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0263818Z triton_mm_278 0.0296 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0264303Z triton_mm_270 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0264770Z triton_mm_271 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0265333Z triton_mm_274 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0265677Z SingleProcess AUTOTUNE benchmarking takes 0.2011 seconds and 0.6415 seconds precompiling for 15 choices 2025-12-04T09:41:13.0266028Z Compiled module path: /tmp/tmpuyrn9hru/66/c66x747yoejfwu2glewx7h3jab2qxz2k5kaaujahxklodqyj4cnu.py 2025-12-04T09:41:13.0266381Z Compiled module path: /tmp/tmpuyrn9hru/y7/cy7cdv3w3mgfm2t6nqc7yr2wtagkyvkt74hbgdknhwoqeeslovzq.py 2025-12-04T09:41:13.0266559Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0266656Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0266791Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0267042Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0268784Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0268866Z graph_break [] 2025-12-04T09:41:13.0268973Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0269147Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0269275Z Autotune Choices Stats: 2025-12-04T09:41:13.0270149Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_302", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0270297Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0270384Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0270490Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0270964Z triton_mm_302 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0271438Z triton_mm_303 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0271907Z triton_mm_305 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0272383Z triton_mm_307 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0272859Z triton_mm_309 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0273318Z triton_mm_300 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0273788Z triton_mm_301 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0274344Z triton_mm_304 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0275079Z triton_mm_306 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0275670Z triton_mm_308 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0276019Z SingleProcess AUTOTUNE benchmarking takes 0.2012 seconds and 0.6271 seconds precompiling for 15 choices 2025-12-04T09:41:13.0276474Z Compiled module path: /tmp/tmpnxfvglo0/mv/cmv5svnrqdgxfpijsujfb6lt3msltpw2rqji7ouqvhdeeykjqzmz.py 2025-12-04T09:41:13.0276832Z Compiled module path: /tmp/tmpnxfvglo0/rx/crxtbd5ejbufzmuyijmfl3i77mx7bc25cjvtl52cepcxot7477r3.py 2025-12-04T09:41:13.0277022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0277153Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0277336Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0277657Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0279469Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0279564Z graph_break [] 2025-12-04T09:41:13.0279671Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0279854Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0280004Z Autotune Choices Stats: 2025-12-04T09:41:13.0280832Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_333", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0280932Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0281031Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0281144Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0281626Z triton_mm_333 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0282106Z triton_mm_335 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0282589Z triton_mm_343 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0283075Z triton_mm_337 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0283541Z triton_mm_332 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0284023Z triton_mm_338 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0284488Z triton_mm_330 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0284971Z triton_mm_331 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0285544Z triton_mm_334 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0286011Z triton_mm_336 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0286367Z SingleProcess AUTOTUNE benchmarking takes 0.4010 seconds and 0.6248 seconds precompiling for 15 choices 2025-12-04T09:41:13.0286731Z Compiled module path: /tmp/tmp9c3250tk/oj/cojlvuknqajncjwhcs6yfkfjjtihtyenckj7m5l37blt326fq55c.py 2025-12-04T09:41:13.0287179Z Compiled module path: /tmp/tmp9c3250tk/st/cst5rwasvqebnwrew6gvackeisqs7qwte54od7766agjws6s3ura.py 2025-12-04T09:41:13.0287366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0287512Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0287658Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0287918Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0289607Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0289695Z graph_break [] 2025-12-04T09:41:13.0289806Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0290037Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0290136Z Autotune Choices Stats: 2025-12-04T09:41:13.0291138Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_367", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0291289Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0291380Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0291502Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0291985Z triton_mm_367 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0292463Z triton_mm_368 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0292956Z triton_mm_369 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0293443Z triton_mm_371 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0293941Z triton_mm_373 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0294419Z triton_mm_363 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0294889Z triton_mm_362 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0295445Z triton_mm_364 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0295922Z triton_mm_365 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0296389Z triton_mm_366 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0296734Z SingleProcess AUTOTUNE benchmarking takes 0.2034 seconds and 0.6302 seconds precompiling for 15 choices 2025-12-04T09:41:13.0297106Z Compiled module path: /tmp/tmpjh1d_rv8/lv/clvgfgovaqoxq6w2a3x3mdsjcbqczglzcgczjlxqwlgcbiom7udy.py 2025-12-04T09:41:13.0297458Z Compiled module path: /tmp/tmpjh1d_rv8/jj/cjjeeuywsjc5aexqz2oy52jb5r72fpvdjcb4gex7fwek3nrj3hsb.py 2025-12-04T09:41:13.0297689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0297793Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0297931Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0298190Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0300063Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0300203Z graph_break [] 2025-12-04T09:41:13.0300316Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0300507Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0300611Z Autotune Choices Stats: 2025-12-04T09:41:13.0301436Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_393", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0301541Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0301630Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0301743Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0302228Z triton_mm_393 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0302706Z triton_mm_396 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0303211Z triton_mm_403 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0303678Z triton_mm_395 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0304154Z triton_mm_401 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0304630Z triton_mm_391 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0305179Z triton_mm_392 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0305654Z triton_mm_394 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0306128Z triton_mm_397 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0306607Z triton_mm_398 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0307020Z SingleProcess AUTOTUNE benchmarking takes 0.2038 seconds and 0.6255 seconds precompiling for 15 choices 2025-12-04T09:41:13.0307391Z Compiled module path: /tmp/tmp23tycbp4/jm/cjmtzqck2innrgpl6fb7ywuwcpt4xhwrp3gneiorpkwkqojqod7x.py 2025-12-04T09:41:13.0307882Z Compiled module path: /tmp/tmp23tycbp4/jx/cjxhng63k4y7iiuaxtdkzcsiabmxi5ytaoafh4mubljd3jqi5aae.py 2025-12-04T09:41:13.0308071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0308171Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0308322Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0308580Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0310364Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0310499Z graph_break [] 2025-12-04T09:41:13.0310621Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0310807Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0310904Z Autotune Choices Stats: 2025-12-04T09:41:13.0311750Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_434", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0311854Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0311947Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0312070Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0312564Z triton_mm_434 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0313054Z triton_mm_427 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0313526Z triton_mm_431 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0314005Z triton_mm_423 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0314526Z triton_mm_421 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0315037Z triton_mm_424 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0315662Z triton_mm_426 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0316188Z triton_mm_430 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0316679Z triton_mm_432 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0317155Z triton_mm_433 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0317559Z SingleProcess AUTOTUNE benchmarking takes 0.2047 seconds and 0.6105 seconds precompiling for 15 choices 2025-12-04T09:41:13.0317931Z Compiled module path: /tmp/tmphd5usaet/dc/cdcdxyzy6cd3jbgwclr23jsuvgkhqh5tbf32wst7ra4nnzfd3rty.py 2025-12-04T09:41:13.0318286Z Compiled module path: /tmp/tmphd5usaet/6k/c6kc65to4c26vezog4ygmrvhsa6qc6xsjj62rc72vinetzwzf6da.py 2025-12-04T09:41:13.0318480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0318583Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0318728Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0318993Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0320814Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0320981Z graph_break [] 2025-12-04T09:41:13.0321090Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0321280Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0321379Z Autotune Choices Stats: 2025-12-04T09:41:13.0322212Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_463", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0322325Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0322416Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0322529Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0323028Z triton_mm_463 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0323502Z triton_mm_450 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0323975Z triton_mm_451 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0324440Z triton_mm_453 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0325004Z triton_mm_464 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0325740Z triton_mm_455 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0326251Z triton_mm_458 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0326737Z triton_mm_460 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0327416Z triton_mm_452 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0327902Z triton_mm_454 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0328300Z SingleProcess AUTOTUNE benchmarking takes 0.2215 seconds and 0.6362 seconds precompiling for 15 choices 2025-12-04T09:41:13.0328674Z Compiled module path: /tmp/tmpc9pvlbau/hy/chya6s2u6d7btbkujfwyicdxobscvkdvifjkht7lz23hxz7lly2d.py 2025-12-04T09:41:13.0329040Z Compiled module path: /tmp/tmpc9pvlbau/jb/cjb6n2psnnfj2lf4fxigzmxslomoqidmvqxghrzqqz2wypsryyuy.py 2025-12-04T09:41:13.0329222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0329328Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0329469Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0329725Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0331464Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0331557Z graph_break [] 2025-12-04T09:41:13.0331673Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0331854Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0331953Z Autotune Choices Stats: 2025-12-04T09:41:13.0332880Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_484", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0332985Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0333079Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0333194Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0333670Z triton_mm_484 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0334155Z triton_mm_485 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0334632Z triton_mm_486 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0335117Z triton_mm_487 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0336299Z triton_mm_488 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0336795Z triton_mm_489 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0337267Z triton_mm_481 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0337789Z triton_mm_482 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0338330Z triton_mm_483 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0338960Z triton_mm_480 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0339313Z SingleProcess AUTOTUNE benchmarking takes 0.2056 seconds and 0.6477 seconds precompiling for 15 choices 2025-12-04T09:41:13.0339735Z Compiled module path: /tmp/tmp7po455mk/zq/czqdrwg6wlfanobk5rsuxzsuswzxqmizjrabbyvkkjoahvmngyue.py 2025-12-04T09:41:13.0340178Z Compiled module path: /tmp/tmp7po455mk/hc/chchsjouzzt5jvril5awri5ugmn3hzvu3i7mr4opswkq5aqbbj7y.py 2025-12-04T09:41:13.0340367Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0340464Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0340607Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0341015Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0342755Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0342852Z graph_break [] 2025-12-04T09:41:13.0342959Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0343145Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0343239Z Autotune Choices Stats: 2025-12-04T09:41:13.0344066Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_521", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0344181Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0344270Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0344386Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0344863Z triton_mm_521 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0345344Z triton_mm_524 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0345825Z triton_mm_517 0.0277 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0346379Z triton_mm_515 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0346903Z triton_mm_520 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0347378Z triton_mm_523 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0347856Z triton_mm_511 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0348325Z triton_mm_512 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0348852Z triton_mm_513 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0349326Z triton_mm_518 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0349670Z SingleProcess AUTOTUNE benchmarking takes 0.2023 seconds and 0.6312 seconds precompiling for 15 choices 2025-12-04T09:41:13.0350037Z Compiled module path: /tmp/tmpe_p5bjio/6o/c6ojtkxpkeubqosthw3efvd63vixnbse24c2gehuwspqfp2fx4at.py 2025-12-04T09:41:13.0350382Z Compiled module path: /tmp/tmpe_p5bjio/ct/cct6e5xbjmvmf7fbjj3pm3amu4sajwis4x3cmls64nm7u3zk74l7.py 2025-12-04T09:41:13.0350561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0350710Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0350849Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0351117Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0352890Z inductor [('triton_bundler_save_kernel', 304), ('benchmarking.InductorBenchmarker.benchmark_gpu', 39), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 24), ('coordesc_tuning_bench', 17), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0352989Z graph_break [] 2025-12-04T09:41:13.0353100Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0353284Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0353392Z Autotune Choices Stats: 2025-12-04T09:41:13.0354234Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_553", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0354331Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0354431Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0354548Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0355046Z triton_mm_553 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0355531Z triton_mm_554 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0356010Z triton_mm_547 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0356568Z triton_mm_548 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0357043Z triton_mm_551 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0357532Z triton_mm_542 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0358001Z triton_mm_541 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0358546Z triton_mm_545 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0359014Z triton_mm_540 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0359481Z triton_mm_543 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0359839Z SingleProcess AUTOTUNE benchmarking takes 0.2014 seconds and 0.6266 seconds precompiling for 15 choices 2025-12-04T09:41:13.0360193Z Compiled module path: /tmp/tmpg8i5cgkd/em/cem6yffkh3my5x7um7rd7ufbeqaw7t6yzo4e75mwl2hys3ne5xbk.py 2025-12-04T09:41:13.0360561Z Compiled module path: /tmp/tmpg8i5cgkd/fy/cfyheiorobaq6ecqx7rnimkff2o3p2venjydoamv4ukkzofqfnc6.py 2025-12-04T09:41:13.0360781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0360882Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0361036Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0361289Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0362979Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0363065Z graph_break [] 2025-12-04T09:41:13.0363177Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0363368Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0363463Z Autotune Choices Stats: 2025-12-04T09:41:13.0364295Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_572", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0364390Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0364480Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0364599Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0365078Z triton_mm_572 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0365557Z triton_mm_573 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0366126Z triton_mm_577 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0366603Z triton_mm_580 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0367139Z triton_mm_581 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0367609Z triton_mm_570 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0368088Z triton_mm_571 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0368605Z triton_mm_574 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0369078Z triton_mm_575 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0369543Z triton_mm_576 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0369888Z SingleProcess AUTOTUNE benchmarking takes 0.2085 seconds and 0.6378 seconds precompiling for 15 choices 2025-12-04T09:41:13.0370269Z Compiled module path: /tmp/tmp5qwmyj31/f4/cf4akhntch3dmtxisx7zye35tofaic5dk6pypiqukvc5hwaop3fe.py 2025-12-04T09:41:13.0370679Z Compiled module path: /tmp/tmp5qwmyj31/ha/chawfc4mtkasifxs6qvh2eotnj3krupriktqki2yomusm5haaai6.py 2025-12-04T09:41:13.0370878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0370976Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0371116Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0371376Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0373054Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0373159Z graph_break [] 2025-12-04T09:41:13.0373268Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0373451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0373553Z Autotune Choices Stats: 2025-12-04T09:41:13.0374374Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_603", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0374477Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0374566Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0374679Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0375157Z triton_mm_603 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0375712Z triton_mm_604 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0376192Z triton_mm_605 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0376667Z triton_mm_606 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0377144Z triton_mm_609 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0377636Z triton_mm_611 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0378167Z triton_mm_613 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0378660Z triton_mm_614 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0379130Z triton_mm_601 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0379606Z triton_mm_602 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0379989Z SingleProcess AUTOTUNE benchmarking takes 0.2008 seconds and 0.6272 seconds precompiling for 15 choices 2025-12-04T09:41:13.0380400Z Compiled module path: /tmp/tmpo9rrm3hb/j2/cj2kdgccny72hbdtda3xzix2me555tucumvq6dcuvocaexfvm3zb.py 2025-12-04T09:41:13.0380766Z Compiled module path: /tmp/tmpo9rrm3hb/4b/c4brg7cx2lwg4dsspsrd2lyeawoavk4cpqnc52xgvpv7ub4xyife.py 2025-12-04T09:41:13.0380951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0381058Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0381199Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0381455Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0383147Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0383245Z graph_break [] 2025-12-04T09:41:13.0383362Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0383542Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0383639Z Autotune Choices Stats: 2025-12-04T09:41:13.0384483Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_639", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0384580Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0384674Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0384787Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0385275Z triton_mm_639 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0385835Z triton_mm_632 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0386310Z triton_mm_638 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0386793Z triton_mm_637 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0387262Z triton_mm_633 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0387753Z triton_mm_643 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0388283Z triton_mm_631 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0388752Z triton_mm_634 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0389228Z triton_mm_635 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0389696Z triton_mm_636 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0390091Z SingleProcess AUTOTUNE benchmarking takes 0.2032 seconds and 0.6400 seconds precompiling for 15 choices 2025-12-04T09:41:13.0390447Z Compiled module path: /tmp/tmpvk_v1nt1/al/calvdf47t4gsdri4v57dyyzt4maptidmnfghvptx324c6ueqz473.py 2025-12-04T09:41:13.0390798Z Compiled module path: /tmp/tmpvk_v1nt1/lu/cluo2f4vnzntzynvmewbgdaorgxwvc4mjp75zmm3fsdv5nestqdm.py 2025-12-04T09:41:13.0390991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0391089Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0391243Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0391498Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0393183Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0393287Z graph_break [] 2025-12-04T09:41:13.0393400Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0393589Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0393685Z Autotune Choices Stats: 2025-12-04T09:41:13.0394510Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_668", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0394620Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0394710Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0394832Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0395419Z triton_mm_668 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0395903Z triton_mm_673 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0396382Z triton_mm_661 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0396855Z triton_mm_662 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0397344Z triton_mm_664 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0397854Z triton_mm_666 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0398337Z triton_mm_660 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0398804Z triton_mm_663 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0399269Z triton_mm_665 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0399785Z triton_mm_669 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0400159Z SingleProcess AUTOTUNE benchmarking takes 0.2085 seconds and 0.6332 seconds precompiling for 15 choices 2025-12-04T09:41:13.0400558Z Compiled module path: /tmp/tmp3o38vp6k/3x/c3xl4taf4zn5lox64lkqtfkt3qjcksh7v5cdlstbqy5a3odjxtz6.py 2025-12-04T09:41:13.0400922Z Compiled module path: /tmp/tmp3o38vp6k/4x/c4xjnodchj6ibvtcnowkrtx3oyfoipw5k4mysuusfoalorvm7uls.py 2025-12-04T09:41:13.0401103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0401207Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0401350Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0401611Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0403298Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0403395Z graph_break [] 2025-12-04T09:41:13.0403508Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0403700Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0403798Z Autotune Choices Stats: 2025-12-04T09:41:13.0404630Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_691", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0404741Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0404833Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0405027Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0405516Z triton_mm_691 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0405989Z triton_mm_692 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0406463Z triton_mm_693 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0406954Z triton_mm_694 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0407528Z triton_mm_695 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0407990Z triton_mm_696 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0408471Z triton_mm_697 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0408945Z triton_mm_698 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0409416Z triton_mm_699 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0410032Z triton_mm_700 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0410381Z SingleProcess AUTOTUNE benchmarking takes 0.2041 seconds and 0.6475 seconds precompiling for 15 choices 2025-12-04T09:41:13.0410747Z Compiled module path: /tmp/tmpcjdheg_x/uy/cuykhl6ore3zfznkrueimqubxt4uk3tpjenrdcqvrtamvvwmfngl.py 2025-12-04T09:41:13.0411093Z Compiled module path: /tmp/tmpcjdheg_x/aq/caqvsnyf5h5bv3f4uhfwimqvm7hf3b6we6pwdlf64itlenqopdlb.py 2025-12-04T09:41:13.0411273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0411384Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0411522Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0411793Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0413477Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0413563Z graph_break [] 2025-12-04T09:41:13.0413680Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0413860Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0413958Z Autotune Choices Stats: 2025-12-04T09:41:13.0414780Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_725", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0414961Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0415060Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0415174Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0415660Z triton_mm_725 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0416147Z triton_mm_731 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0416617Z triton_mm_723 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0417096Z triton_mm_726 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0417611Z triton_mm_728 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0418097Z triton_mm_729 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0418575Z triton_mm_730 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0419063Z triton_mm_732 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0419580Z triton_mm_733 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0420061Z triton_mm_734 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0420424Z SingleProcess AUTOTUNE benchmarking takes 0.2070 seconds and 0.6359 seconds precompiling for 15 choices 2025-12-04T09:41:13.0420781Z Compiled module path: /tmp/tmpcri9z8ii/ll/clleccci72holl57y6qdvpjw4nqe2v3o33e3b4dbi52l4yonc2u4.py 2025-12-04T09:41:13.0421159Z Compiled module path: /tmp/tmpcri9z8ii/nt/cntotzvaz6lwon2hethxzh4kfya2upvsk7elnoqjviiakmz27h65.py 2025-12-04T09:41:13.0421342Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0421448Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0421598Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0421856Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0423553Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0423646Z graph_break [] 2025-12-04T09:41:13.0423756Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0423951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0424052Z Autotune Choices Stats: 2025-12-04T09:41:13.0424976Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_758", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.027744000777602196, "best_triton_pos": 0} 2025-12-04T09:41:13.0425082Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0425173Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0425298Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0425778Z triton_mm_758 0.0277 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0426262Z triton_mm_753 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0426739Z triton_mm_755 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0427258Z triton_mm_759 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0427753Z triton_mm_760 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0428224Z triton_mm_761 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0428715Z triton_mm_764 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0429230Z triton_mm_752 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0429716Z triton_mm_751 0.0297 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0430234Z triton_mm_754 0.0297 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0430586Z SingleProcess AUTOTUNE benchmarking takes 0.2023 seconds and 0.6256 seconds precompiling for 15 choices 2025-12-04T09:41:13.0430944Z Compiled module path: /tmp/tmp_crk5itx/7h/c7hlonnyw5ynm5th3n7w5hspkcrh4z7yqrc5z26yajh3wiscnl4h.py 2025-12-04T09:41:13.0431292Z Compiled module path: /tmp/tmp_crk5itx/er/cerbi2equovlvhddb3tgbmag3wpo2vfegtrxofu5awfig5xspie5.py 2025-12-04T09:41:13.0431488Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0431591Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0431727Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0431989Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0433662Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0433760Z graph_break [] 2025-12-04T09:41:13.0433866Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0434047Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0434152Z Autotune Choices Stats: 2025-12-04T09:41:13.0435078Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_781", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.02969600073993206, "best_triton_pos": 0} 2025-12-04T09:41:13.0435181Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0435526Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0435683Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0436220Z triton_mm_781 0.0297 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0436701Z triton_mm_794 0.0297 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0437266Z triton_mm_780 0.0307 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0437734Z triton_mm_787 0.0307 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0438203Z triton_mm_789 0.0307 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0438678Z triton_mm_790 0.0307 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0439148Z triton_mm_792 0.0307 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0439704Z triton_mm_788 0.0308 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0440168Z triton_mm_782 0.0317 ms 93.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0440637Z triton_mm_783 0.0317 ms 93.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0440976Z SingleProcess AUTOTUNE benchmarking takes 0.2089 seconds and 0.6275 seconds precompiling for 15 choices 2025-12-04T09:41:13.0441335Z Compiled module path: /tmp/tmpat625dol/5j/c5jcehve6psqmluc2fkhixgt3hzwttzv5cbgxyzh3di3c76stb5j.py 2025-12-04T09:41:13.0441699Z Compiled module path: /tmp/tmpat625dol/kn/cknktcrdvm4elzlwkhq5s63lc7s4ucgqeicjmy6tacyszdhzz5a3.py 2025-12-04T09:41:13.0441887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0441994Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0442134Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0442386Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0444070Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0444162Z graph_break [] 2025-12-04T09:41:13.0444279Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0444562Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0444663Z Autotune Choices Stats: 2025-12-04T09:41:13.0445489Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_810", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0445585Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0445683Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0445793Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0446267Z triton_mm_810 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0446751Z triton_mm_812 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0447315Z triton_mm_813 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0447798Z triton_mm_815 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0448267Z triton_mm_816 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0448741Z triton_mm_817 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0449279Z triton_mm_819 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0449760Z triton_mm_820 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0450238Z triton_mm_821 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0450708Z triton_mm_811 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0451059Z SingleProcess AUTOTUNE benchmarking takes 0.2007 seconds and 0.6373 seconds precompiling for 15 choices 2025-12-04T09:41:13.0451423Z Compiled module path: /tmp/tmp1pvy590j/wz/cwz3w7hwh6gvj22bftrdtw7bfbogxwogh7rm4vahl7d4y2ddh2u5.py 2025-12-04T09:41:13.0451785Z Compiled module path: /tmp/tmp1pvy590j/y2/cy2v3dpgqco7sv23sigs3qjjosno5eomtrgfwlewtcvvuagtado3.py 2025-12-04T09:41:13.0451973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0452073Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0452209Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0452468Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0454147Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0454248Z graph_break [] 2025-12-04T09:41:13.0454436Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0454628Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0454726Z Autotune Choices Stats: 2025-12-04T09:41:13.0455548Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_846", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0455654Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0455741Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0455851Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0456338Z triton_mm_846 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0456873Z triton_mm_847 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0457354Z triton_mm_849 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0457842Z triton_mm_853 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0458322Z triton_mm_848 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0458833Z triton_mm_850 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0459321Z triton_mm_851 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0459799Z triton_mm_852 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0460277Z triton_mm_854 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0460753Z triton_mm_845 0.0308 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0461099Z SingleProcess AUTOTUNE benchmarking takes 0.2058 seconds and 0.6435 seconds precompiling for 15 choices 2025-12-04T09:41:13.0461472Z Compiled module path: /tmp/tmpwhwf0ibm/7v/c7v6bgeyvxnrvpshs2r4f3wzf5esyopbsdidn2nmcac4x5wjjwnx.py 2025-12-04T09:41:13.0461838Z Compiled module path: /tmp/tmpwhwf0ibm/tf/ctf7xbdusnrfspyrmou3lxcbkogwc2fspbbmznenz7ifzfbfs456.py 2025-12-04T09:41:13.0462017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0462117Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0462254Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0462506Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0464267Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0464362Z graph_break [] 2025-12-04T09:41:13.0464478Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0464660Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0464759Z Autotune Choices Stats: 2025-12-04T09:41:13.0465585Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_871", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0465679Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0465778Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0465889Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0466415Z triton_mm_871 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0466901Z triton_mm_872 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0467374Z triton_mm_873 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0467845Z triton_mm_874 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0468320Z triton_mm_875 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0468867Z triton_mm_876 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0469345Z triton_mm_877 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0469827Z triton_mm_878 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0470308Z triton_mm_879 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0470791Z triton_mm_880 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0471150Z SingleProcess AUTOTUNE benchmarking takes 0.2002 seconds and 0.6285 seconds precompiling for 15 choices 2025-12-04T09:41:13.0471512Z Compiled module path: /tmp/tmpmwjjx73y/wb/cwbsd56cyejfu6yitythbmga3wllzq4qb3mhdxxhxv7igpsk6rkf.py 2025-12-04T09:41:13.0471879Z Compiled module path: /tmp/tmpmwjjx73y/xu/cxuzsholowpf77keyfoipo3tmz3nako2i2tsxuhkjqzbk2unl3bz.py 2025-12-04T09:41:13.0472061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0472157Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0472298Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0472548Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0474315Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0474407Z graph_break [] 2025-12-04T09:41:13.0474515Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0474701Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0474800Z Autotune Choices Stats: 2025-12-04T09:41:13.0475628Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_902", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0475769Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0475862Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0475978Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0476462Z triton_mm_902 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0476938Z triton_mm_903 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0477418Z triton_mm_905 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0477899Z triton_mm_907 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0478426Z triton_mm_908 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0478904Z triton_mm_909 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0479393Z triton_mm_911 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0479981Z triton_mm_913 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0480572Z triton_mm_900 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0481134Z triton_mm_901 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0481526Z SingleProcess AUTOTUNE benchmarking takes 0.2012 seconds and 0.7632 seconds precompiling for 15 choices 2025-12-04T09:41:13.0481934Z Compiled module path: /tmp/tmphyk566_q/yu/cyuea4r4os7pugqxnksie3dv6pq7fo2ns77dqcfthvdvgf4r52j3.py 2025-12-04T09:41:13.0482344Z Compiled module path: /tmp/tmphyk566_q/ly/clymrvghrylu2ct5p7svacigybspd2ksvswnqbpvhrgbdyv7v3tn.py 2025-12-04T09:41:13.0482548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0482648Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0482792Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0483085Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0485244Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0485338Z graph_break [] 2025-12-04T09:41:13.0485454Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0485655Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0485766Z Autotune Choices Stats: 2025-12-04T09:41:13.0486778Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_939", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0486936Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0487097Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0487210Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0487796Z triton_mm_939 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0488357Z triton_mm_931 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0488921Z triton_mm_932 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0489432Z triton_mm_933 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0489917Z triton_mm_935 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0490400Z triton_mm_937 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0490875Z triton_mm_941 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0491355Z triton_mm_940 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0491831Z triton_mm_944 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0492315Z triton_mm_938 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0492662Z SingleProcess AUTOTUNE benchmarking takes 0.2030 seconds and 0.6686 seconds precompiling for 15 choices 2025-12-04T09:41:13.0493021Z Compiled module path: /tmp/tmp4cvs758s/n6/cn6tmzlh3kkhsuhuer55fd5jb2spgvj2yyqbmx3rllxtqyhuj3t2.py 2025-12-04T09:41:13.0493385Z Compiled module path: /tmp/tmp4cvs758s/6r/c6r375atdd4kvestjgimuafqdjuwpz5lxwv7wbi4jwlyczo6vzzt.py 2025-12-04T09:41:13.0493564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0493663Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0493815Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0494068Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0495915Z inductor [('triton_bundler_save_kernel', 288), ('benchmarking.InductorBenchmarker.benchmark_gpu', 37), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 22), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('coordesc_tuning_bench', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0496003Z graph_break [] 2025-12-04T09:41:13.0496117Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0496299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0496395Z Autotune Choices Stats: 2025-12-04T09:41:13.0497285Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_974", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0497381Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0497469Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0497594Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0498080Z triton_mm_974 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0498566Z triton_mm_961 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0499082Z triton_mm_969 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0499576Z triton_mm_967 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0500048Z triton_mm_962 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0500519Z triton_mm_963 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0500999Z triton_mm_965 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0501471Z triton_mm_966 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0501969Z triton_mm_968 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0502442Z triton_mm_970 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0502795Z SingleProcess AUTOTUNE benchmarking takes 0.2037 seconds and 0.6293 seconds precompiling for 15 choices 2025-12-04T09:41:13.0503163Z Compiled module path: /tmp/tmpzrw3psn5/ex/cexdxarjh6qlkmrh5ex2jzrhiw737frfohgvjrzxgquhfv5g2o4l.py 2025-12-04T09:41:13.0503528Z Compiled module path: /tmp/tmpzrw3psn5/3v/c3vwhsbwrmtggganhnbbkwvtdwpm4qfdu33rqhqhbdczjvs5z6rs.py 2025-12-04T09:41:13.0503828Z ______ BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code _______ 2025-12-04T09:41:13.0503941Z Traceback (most recent call last): 2025-12-04T09:41:13.0504445Z File "/var/lib/jenkins/workspace/test/inductor/test_benchmark_fusion.py", line 303, in test_equivalent_template_code 2025-12-04T09:41:13.0504647Z ).check("" if config.cpp_wrapper else "return").run(out_code[0]) 2025-12-04T09:41:13.0504909Z RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0505012Z Searched string: 2025-12-04T09:41:13.0505126Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0505233Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0505418Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0505665Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0505768Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0505998Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0506150Z del arg0_1 2025-12-04T09:41:13.0506254Z del arg2_1 2025-12-04T09:41:13.0506365Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0506632Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0506744Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0506949Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0507038Z del arg1_1 2025-12-04T09:41:13.0507136Z return (buf1, ) 2025-12-04T09:41:13.0507142Z 2025-12-04T09:41:13.0507249Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0507337Z call = runner.call 2025-12-04T09:41:13.0507518Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0507523Z 2025-12-04T09:41:13.0507527Z 2025-12-04T09:41:13.0507679Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0507883Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0508040Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0508260Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0508468Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0508674Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0508797Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0508970Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0508975Z 2025-12-04T09:41:13.0508979Z 2025-12-04T09:41:13.0509073Z if __name__ == "__main__": 2025-12-04T09:41:13.0509295Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0509467Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0509596Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0509608Z 2025-12-04T09:41:13.0509612Z 2025-12-04T09:41:13.0509801Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0510508Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0510513Z 2025-12-04T09:41:13.0510744Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0510926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0511036Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0511178Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0512944Z inductor [('triton_bundler_save_kernel', 216), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('async_compile_cache_miss', 7), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0513222Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0513309Z graph_break [] 2025-12-04T09:41:13.0513423Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0513601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0514916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py:3433: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage() 2025-12-04T09:41:13.0515084Z current_size = base.storage().size() 2025-12-04T09:41:13.0515186Z Autotune Choices Stats: 2025-12-04T09:41:13.0516181Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_10", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0516288Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0516376Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0516534Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0517057Z triton_mm_10 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0517624Z triton_mm_9 0.0277 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0518169Z triton_mm_11 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0518732Z triton_mm_13 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0519269Z triton_mm_14 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0519783Z triton_mm_12 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0520265Z triton_mm_1 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0520740Z triton_mm_2 0.0317 ms 87.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0521206Z triton_mm_6 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0521682Z triton_mm_7 0.0317 ms 87.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0522028Z SingleProcess AUTOTUNE benchmarking takes 0.1729 seconds and 1.0713 seconds precompiling for 15 choices 2025-12-04T09:41:13.0522405Z Compiled module path: /tmp/tmpgejy12ej/qv/cqvwpzgbltbzt3unpdzscv64a2elpufsmhey6btnq73pj2f4vaa7.py 2025-12-04T09:41:13.0522768Z Compiled module path: /tmp/tmpgejy12ej/jj/cjj4tsxlplxvzz5g4jfphbz5jsxl7b3fd5mxbhxre5lcbuh3vmrc.py 2025-12-04T09:41:13.0522961Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0523149Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0523289Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0523556Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0525246Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0525381Z graph_break [] 2025-12-04T09:41:13.0525488Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0525676Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0525775Z Autotune Choices Stats: 2025-12-04T09:41:13.0526600Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_30", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0526706Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0532279Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0532413Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0532906Z triton_mm_30 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0533469Z triton_mm_32 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0533959Z triton_mm_33 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0534422Z triton_mm_34 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0534955Z triton_mm_35 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0535853Z triton_mm_36 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0536416Z triton_mm_37 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0536992Z triton_mm_38 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0537467Z triton_mm_39 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0538030Z triton_mm_40 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0538379Z SingleProcess AUTOTUNE benchmarking takes 0.2013 seconds and 0.7407 seconds precompiling for 15 choices 2025-12-04T09:41:13.0538834Z Compiled module path: /tmp/tmpu9odmkxw/ai/cai5yet7zvfv3ry4tqny7aptct3cbabbgknxxt3otvygyekzraji.py 2025-12-04T09:41:13.0539357Z Compiled module path: /tmp/tmpu9odmkxw/nq/cnqm37sxiqmqocnfb7wdcdqxnd56ra26fhpzusixi763umecdh56.py 2025-12-04T09:41:13.0539612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0539748Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0539901Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0540195Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0542144Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 36), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 21), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 14), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0542330Z graph_break [] 2025-12-04T09:41:13.0542447Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0542630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0542726Z Autotune Choices Stats: 2025-12-04T09:41:13.0543576Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_73", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0543675Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0543770Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0543882Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0544423Z triton_mm_73 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0544908Z triton_mm_62 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0545372Z triton_mm_65 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0545858Z triton_mm_71 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0546324Z triton_mm_61 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0546794Z triton_mm_63 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0547277Z triton_mm_60 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0547736Z triton_mm_64 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0548218Z triton_mm_67 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0548684Z triton_mm_68 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0549034Z SingleProcess AUTOTUNE benchmarking takes 0.2006 seconds and 0.6267 seconds precompiling for 15 choices 2025-12-04T09:41:13.0549517Z Compiled module path: /tmp/tmpqun5dtux/q7/cq7yaif2wknvncdje2hbdwcngroryoj2mtkk4dkwdkx45aat6hm7.py 2025-12-04T09:41:13.0549881Z Compiled module path: /tmp/tmpqun5dtux/y6/cy66mazx6rmemwkg6jecmgbfmpsvz6cfz5y45krjnl3azaph6y7t.py 2025-12-04T09:41:13.0550069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0550167Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0550308Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0550572Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0552271Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0552415Z graph_break [] 2025-12-04T09:41:13.0552523Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0552713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0552811Z Autotune Choices Stats: 2025-12-04T09:41:13.0553643Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_91", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0553748Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0553881Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0553995Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0554491Z triton_mm_91 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0554967Z triton_mm_95 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0555445Z triton_mm_90 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0555906Z triton_mm_92 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0556370Z triton_mm_93 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0556845Z triton_mm_96 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0557322Z triton_mm_97 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0557801Z triton_mm_98 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0558271Z triton_mm_94 0.0307 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0558756Z triton_mm_100 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0559188Z SingleProcess AUTOTUNE benchmarking takes 0.2153 seconds and 0.6426 seconds precompiling for 15 choices 2025-12-04T09:41:13.0559563Z Compiled module path: /tmp/tmp2v_kzp9a/fj/cfjz3b7ajfatqwomwryerntjnvk6kd5wysrcdign42gjpmvqkh3t.py 2025-12-04T09:41:13.0559926Z Compiled module path: /tmp/tmp2v_kzp9a/57/c57neznqb6ba2jtzlgbchzekwlmukri7wpo65rmsqusn7t6xlfxx.py 2025-12-04T09:41:13.0560134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0560245Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0560384Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0560635Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0562426Z inductor [('triton_bundler_save_kernel', 288), ('benchmarking.InductorBenchmarker.benchmark_gpu', 38), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 23), ('coordesc_tuning_bench', 16), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0562559Z graph_break [] 2025-12-04T09:41:13.0562669Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0562851Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0562945Z Autotune Choices Stats: 2025-12-04T09:41:13.0563795Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_134", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0563934Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0564031Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0564145Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0564633Z triton_mm_134 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0565120Z triton_mm_121 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0565589Z triton_mm_122 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0566070Z triton_mm_124 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0566546Z triton_mm_125 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0567131Z triton_mm_128 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0567608Z triton_mm_129 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0568081Z triton_mm_130 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0568561Z triton_mm_131 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0569117Z triton_mm_120 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0569471Z SingleProcess AUTOTUNE benchmarking takes 0.2007 seconds and 0.6630 seconds precompiling for 15 choices 2025-12-04T09:41:13.0569836Z Compiled module path: /tmp/tmp2w3f9rlp/jd/cjdbzzfn2s4jgc6gkasctk4suc4halinn3hncy6u26d22lzxdphd.py 2025-12-04T09:41:13.0570206Z Compiled module path: /tmp/tmp2w3f9rlp/4s/c4sivzup5qddnp7nobqsap2nr2ndjcdy3tashkjs4q3gubdqirss.py 2025-12-04T09:41:13.0570390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0570485Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0570634Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0570892Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0572719Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0572813Z graph_break [] 2025-12-04T09:41:13.0572927Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0573118Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0573212Z Autotune Choices Stats: 2025-12-04T09:41:13.0574049Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_158", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0574203Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0574296Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0574419Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0574900Z triton_mm_158 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0575382Z triton_mm_159 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0575861Z triton_mm_161 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0576332Z triton_mm_150 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0576821Z triton_mm_151 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0577298Z triton_mm_153 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0577779Z triton_mm_154 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0578245Z triton_mm_155 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0578798Z triton_mm_156 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0579277Z triton_mm_157 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0579626Z SingleProcess AUTOTUNE benchmarking takes 0.2054 seconds and 0.6100 seconds precompiling for 15 choices 2025-12-04T09:41:13.0580000Z Compiled module path: /tmp/tmpwi6t5aeb/gn/cgneadkz2tnedarmlzhqm4gn55whqeexaln2fzmkic7oxhyres4r.py 2025-12-04T09:41:13.0580355Z Compiled module path: /tmp/tmpwi6t5aeb/gy/cgyxpcavcgmmp26t6h56dechg7qqid3n7dyu5q6u5h3b6c7fepoj.py 2025-12-04T09:41:13.0580533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0580639Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0580782Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0581084Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0582776Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0582870Z graph_break [] 2025-12-04T09:41:13.0582979Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0583157Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0583298Z Autotune Choices Stats: 2025-12-04T09:41:13.0584131Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_186", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0584234Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0584330Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0584441Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0584920Z triton_mm_186 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0585397Z triton_mm_193 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0585871Z triton_mm_182 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0586369Z triton_mm_194 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0586835Z triton_mm_184 0.0308 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0587313Z triton_mm_187 0.0317 ms 90.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0587779Z triton_mm_185 0.0317 ms 90.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0588253Z triton_mm_180 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0588830Z triton_mm_181 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0589300Z triton_mm_183 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0589656Z SingleProcess AUTOTUNE benchmarking takes 0.2087 seconds and 0.6406 seconds precompiling for 15 choices 2025-12-04T09:41:13.0590023Z Compiled module path: /tmp/tmpk78l5vwh/kk/ckk35oaicdyefn5yev5gn7hgiqa5qquexctvfdcqebmw6q6whm6z.py 2025-12-04T09:41:13.0590387Z Compiled module path: /tmp/tmpk78l5vwh/hm/chmuaf5ovlicrcx6s7zj4pujz62mn7yxmmkafqcdbwqtfoym5ys5.py 2025-12-04T09:41:13.0590568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0590704Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0590846Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0591103Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0592794Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0592877Z graph_break [] 2025-12-04T09:41:13.0592984Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0593210Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0593310Z Autotune Choices Stats: 2025-12-04T09:41:13.0594149Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_220", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0594247Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0594336Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0594452Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0594931Z triton_mm_220 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0595409Z triton_mm_211 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0595887Z triton_mm_212 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0596348Z triton_mm_214 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0596818Z triton_mm_215 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0597286Z triton_mm_216 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0597759Z triton_mm_217 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0598311Z triton_mm_218 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0598786Z triton_mm_219 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0599253Z triton_mm_221 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0599592Z SingleProcess AUTOTUNE benchmarking takes 0.1998 seconds and 0.6553 seconds precompiling for 15 choices 2025-12-04T09:41:13.0599965Z Compiled module path: /tmp/tmpwuxhdg2k/ip/cipioi7ficmn3ww6fbrd2oyul2gc6oxnifncerak7mzlxvpbihdc.py 2025-12-04T09:41:13.0600381Z Compiled module path: /tmp/tmpwuxhdg2k/tq/ctqm7bgwweclnxyitnuulxziaydok5e5vcjg4prplje6cyryh3p3.py 2025-12-04T09:41:13.0600611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0600715Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0600853Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0601114Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0602796Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0602926Z graph_break [] 2025-12-04T09:41:13.0603037Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0603214Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0603319Z Autotune Choices Stats: 2025-12-04T09:41:13.0604149Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_250", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0604253Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0604340Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0604450Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0604934Z triton_mm_250 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0605405Z triton_mm_245 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0605892Z triton_mm_249 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0606367Z triton_mm_251 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0606835Z triton_mm_243 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0607378Z triton_mm_241 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0607847Z triton_mm_240 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0608405Z triton_mm_242 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0608876Z triton_mm_244 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0609350Z triton_mm_246 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0609693Z SingleProcess AUTOTUNE benchmarking takes 0.2009 seconds and 0.6388 seconds precompiling for 15 choices 2025-12-04T09:41:13.0610064Z Compiled module path: /tmp/tmpv4covkzm/xp/cxpol2uuscxtvrr3qpm7ebdm7l5updopjmnk75hpevamiwzj65o7.py 2025-12-04T09:41:13.0610464Z Compiled module path: /tmp/tmpv4covkzm/rn/crnctdx35xmvga3ti6xdrde3d3i7bt47oqtpexetsy437yfgpmyy.py 2025-12-04T09:41:13.0610653Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0610758Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0610896Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0611146Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0612829Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0612999Z graph_break [] 2025-12-04T09:41:13.0613115Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0613296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0613390Z Autotune Choices Stats: 2025-12-04T09:41:13.0614217Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_276", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.02768000029027462, "best_triton_pos": 0} 2025-12-04T09:41:13.0614314Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0614409Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0614520Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0614991Z triton_mm_276 0.0277 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0615471Z triton_mm_272 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0615933Z triton_mm_273 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0616413Z triton_mm_277 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0616890Z triton_mm_280 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0617365Z triton_mm_281 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0617929Z triton_mm_278 0.0296 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0618397Z triton_mm_270 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0618870Z triton_mm_271 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0619338Z triton_mm_274 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0619689Z SingleProcess AUTOTUNE benchmarking takes 0.2011 seconds and 0.6415 seconds precompiling for 15 choices 2025-12-04T09:41:13.0620084Z Compiled module path: /tmp/tmpuyrn9hru/66/c66x747yoejfwu2glewx7h3jab2qxz2k5kaaujahxklodqyj4cnu.py 2025-12-04T09:41:13.0620440Z Compiled module path: /tmp/tmpuyrn9hru/y7/cy7cdv3w3mgfm2t6nqc7yr2wtagkyvkt74hbgdknhwoqeeslovzq.py 2025-12-04T09:41:13.0620622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0620722Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0620862Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0621114Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0622804Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0622966Z graph_break [] 2025-12-04T09:41:13.0623072Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0623255Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0623348Z Autotune Choices Stats: 2025-12-04T09:41:13.0624166Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_302", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0624265Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0624353Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0624475Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0624952Z triton_mm_302 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0625423Z triton_mm_303 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0625893Z triton_mm_305 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0626364Z triton_mm_307 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0626978Z triton_mm_309 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0627546Z triton_mm_300 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0628117Z triton_mm_301 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0628639Z triton_mm_304 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0629137Z triton_mm_306 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0629708Z triton_mm_308 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0630862Z SingleProcess AUTOTUNE benchmarking takes 0.2012 seconds and 0.6271 seconds precompiling for 15 choices 2025-12-04T09:41:13.0631266Z Compiled module path: /tmp/tmpnxfvglo0/mv/cmv5svnrqdgxfpijsujfb6lt3msltpw2rqji7ouqvhdeeykjqzmz.py 2025-12-04T09:41:13.0631672Z Compiled module path: /tmp/tmpnxfvglo0/rx/crxtbd5ejbufzmuyijmfl3i77mx7bc25cjvtl52cepcxot7477r3.py 2025-12-04T09:41:13.0631855Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0631963Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0632112Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0632434Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0634237Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0634392Z graph_break [] 2025-12-04T09:41:13.0634503Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0634683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0634783Z Autotune Choices Stats: 2025-12-04T09:41:13.0635970Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_333", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0636077Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0636175Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0636287Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0636772Z triton_mm_333 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0637243Z triton_mm_335 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0637733Z triton_mm_343 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0638219Z triton_mm_337 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0638688Z triton_mm_332 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0639299Z triton_mm_338 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0639770Z triton_mm_330 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0640248Z triton_mm_331 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0640716Z triton_mm_334 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0641251Z triton_mm_336 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0641602Z SingleProcess AUTOTUNE benchmarking takes 0.4010 seconds and 0.6248 seconds precompiling for 15 choices 2025-12-04T09:41:13.0641964Z Compiled module path: /tmp/tmp9c3250tk/oj/cojlvuknqajncjwhcs6yfkfjjtihtyenckj7m5l37blt326fq55c.py 2025-12-04T09:41:13.0642321Z Compiled module path: /tmp/tmp9c3250tk/st/cst5rwasvqebnwrew6gvackeisqs7qwte54od7766agjws6s3ura.py 2025-12-04T09:41:13.0642499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0642600Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0642746Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0643006Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0644764Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0644855Z graph_break [] 2025-12-04T09:41:13.0644963Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0645147Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0645246Z Autotune Choices Stats: 2025-12-04T09:41:13.0646095Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_367", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0646197Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0646291Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0646408Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0646890Z triton_mm_367 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0647427Z triton_mm_368 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0647905Z triton_mm_369 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0648384Z triton_mm_371 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0648959Z triton_mm_373 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0649424Z triton_mm_363 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0649892Z triton_mm_362 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0650356Z triton_mm_364 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0650833Z triton_mm_365 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0651341Z triton_mm_366 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0651683Z SingleProcess AUTOTUNE benchmarking takes 0.2034 seconds and 0.6302 seconds precompiling for 15 choices 2025-12-04T09:41:13.0652038Z Compiled module path: /tmp/tmpjh1d_rv8/lv/clvgfgovaqoxq6w2a3x3mdsjcbqczglzcgczjlxqwlgcbiom7udy.py 2025-12-04T09:41:13.0652381Z Compiled module path: /tmp/tmpjh1d_rv8/jj/cjjeeuywsjc5aexqz2oy52jb5r72fpvdjcb4gex7fwek3nrj3hsb.py 2025-12-04T09:41:13.0652571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0652669Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0652808Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0653112Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0654807Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0654894Z graph_break [] 2025-12-04T09:41:13.0655002Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0655183Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0655282Z Autotune Choices Stats: 2025-12-04T09:41:13.0656113Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_393", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0656220Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0656306Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0656418Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0656895Z triton_mm_393 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0657363Z triton_mm_396 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0657850Z triton_mm_403 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0658398Z triton_mm_395 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0658872Z triton_mm_401 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0659358Z triton_mm_391 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0659824Z triton_mm_392 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0660345Z triton_mm_394 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0660919Z triton_mm_397 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0661393Z triton_mm_398 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0661743Z SingleProcess AUTOTUNE benchmarking takes 0.2038 seconds and 0.6255 seconds precompiling for 15 choices 2025-12-04T09:41:13.0662114Z Compiled module path: /tmp/tmp23tycbp4/jm/cjmtzqck2innrgpl6fb7ywuwcpt4xhwrp3gneiorpkwkqojqod7x.py 2025-12-04T09:41:13.0662472Z Compiled module path: /tmp/tmp23tycbp4/jx/cjxhng63k4y7iiuaxtdkzcsiabmxi5ytaoafh4mubljd3jqi5aae.py 2025-12-04T09:41:13.0662653Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0662790Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0662939Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0663194Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0664978Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0665062Z graph_break [] 2025-12-04T09:41:13.0665173Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0665361Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0665455Z Autotune Choices Stats: 2025-12-04T09:41:13.0666306Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_434", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0666400Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0666496Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0666603Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0667085Z triton_mm_434 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0667566Z triton_mm_427 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0668118Z triton_mm_431 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0668591Z triton_mm_423 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0669061Z triton_mm_421 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0669529Z triton_mm_424 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0670006Z triton_mm_426 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0670530Z triton_mm_430 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0671008Z triton_mm_432 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0671483Z triton_mm_433 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0671821Z SingleProcess AUTOTUNE benchmarking takes 0.2047 seconds and 0.6105 seconds precompiling for 15 choices 2025-12-04T09:41:13.0672185Z Compiled module path: /tmp/tmphd5usaet/dc/cdcdxyzy6cd3jbgwclr23jsuvgkhqh5tbf32wst7ra4nnzfd3rty.py 2025-12-04T09:41:13.0672572Z Compiled module path: /tmp/tmphd5usaet/6k/c6kc65to4c26vezog4ygmrvhsa6qc6xsjj62rc72vinetzwzf6da.py 2025-12-04T09:41:13.0672760Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0672861Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0673000Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0673254Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0675027Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 33), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 18), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 11), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0675123Z graph_break [] 2025-12-04T09:41:13.0675231Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0675414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0675516Z Autotune Choices Stats: 2025-12-04T09:41:13.0676359Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_463", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0676456Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0676540Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0676647Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0677136Z triton_mm_463 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0677605Z triton_mm_450 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0678166Z triton_mm_451 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0678632Z triton_mm_453 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0679113Z triton_mm_464 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0679575Z triton_mm_455 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0680114Z triton_mm_458 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0680620Z triton_mm_460 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0681086Z triton_mm_452 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0681551Z triton_mm_454 0.0317 ms 90.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0681891Z SingleProcess AUTOTUNE benchmarking takes 0.2215 seconds and 0.6362 seconds precompiling for 15 choices 2025-12-04T09:41:13.0682298Z Compiled module path: /tmp/tmpc9pvlbau/hy/chya6s2u6d7btbkujfwyicdxobscvkdvifjkht7lz23hxz7lly2d.py 2025-12-04T09:41:13.0682669Z Compiled module path: /tmp/tmpc9pvlbau/jb/cjb6n2psnnfj2lf4fxigzmxslomoqidmvqxghrzqqz2wypsryyuy.py 2025-12-04T09:41:13.0682847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0682945Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0683079Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0683331Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0685034Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0685125Z graph_break [] 2025-12-04T09:41:13.0685241Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0685420Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0685522Z Autotune Choices Stats: 2025-12-04T09:41:13.0686343Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_484", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0686436Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0686529Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0686634Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0687156Z triton_mm_484 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0687726Z triton_mm_485 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0688195Z triton_mm_486 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0688674Z triton_mm_487 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0689145Z triton_mm_488 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0689631Z triton_mm_489 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0690197Z triton_mm_481 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0690658Z triton_mm_482 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0691129Z triton_mm_483 0.0297 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0691596Z triton_mm_480 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0691985Z SingleProcess AUTOTUNE benchmarking takes 0.2056 seconds and 0.6477 seconds precompiling for 15 choices 2025-12-04T09:41:13.0692356Z Compiled module path: /tmp/tmp7po455mk/zq/czqdrwg6wlfanobk5rsuxzsuswzxqmizjrabbyvkkjoahvmngyue.py 2025-12-04T09:41:13.0692711Z Compiled module path: /tmp/tmp7po455mk/hc/chchsjouzzt5jvril5awri5ugmn3hzvu3i7mr4opswkq5aqbbj7y.py 2025-12-04T09:41:13.0692888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0692981Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0693118Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0693369Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0695060Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0695149Z graph_break [] 2025-12-04T09:41:13.0695254Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0695434Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0695526Z Autotune Choices Stats: 2025-12-04T09:41:13.0696354Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_521", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0696451Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0696539Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0696649Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0697242Z triton_mm_521 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0697722Z triton_mm_524 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0698198Z triton_mm_517 0.0277 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0698663Z triton_mm_515 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0699141Z triton_mm_520 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0699661Z triton_mm_523 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0700183Z triton_mm_511 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0700651Z triton_mm_512 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0701114Z triton_mm_513 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0701589Z triton_mm_518 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0701983Z SingleProcess AUTOTUNE benchmarking takes 0.2023 seconds and 0.6312 seconds precompiling for 15 choices 2025-12-04T09:41:13.0702340Z Compiled module path: /tmp/tmpe_p5bjio/6o/c6ojtkxpkeubqosthw3efvd63vixnbse24c2gehuwspqfp2fx4at.py 2025-12-04T09:41:13.0702683Z Compiled module path: /tmp/tmpe_p5bjio/ct/cct6e5xbjmvmf7fbjj3pm3amu4sajwis4x3cmls64nm7u3zk74l7.py 2025-12-04T09:41:13.0702859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0702957Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0703091Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0703343Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0705121Z inductor [('triton_bundler_save_kernel', 304), ('benchmarking.InductorBenchmarker.benchmark_gpu', 39), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 24), ('coordesc_tuning_bench', 17), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0705211Z graph_break [] 2025-12-04T09:41:13.0705315Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0705492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0705589Z Autotune Choices Stats: 2025-12-04T09:41:13.0706416Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_553", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0706523Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0706609Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0706799Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0707289Z triton_mm_553 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0707766Z triton_mm_554 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0708242Z triton_mm_547 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0708710Z triton_mm_548 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0709229Z triton_mm_551 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0709695Z triton_mm_542 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0710163Z triton_mm_541 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0710631Z triton_mm_545 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0711092Z triton_mm_540 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0711611Z triton_mm_543 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0711956Z SingleProcess AUTOTUNE benchmarking takes 0.2014 seconds and 0.6266 seconds precompiling for 15 choices 2025-12-04T09:41:13.0712308Z Compiled module path: /tmp/tmpg8i5cgkd/em/cem6yffkh3my5x7um7rd7ufbeqaw7t6yzo4e75mwl2hys3ne5xbk.py 2025-12-04T09:41:13.0712674Z Compiled module path: /tmp/tmpg8i5cgkd/fy/cfyheiorobaq6ecqx7rnimkff2o3p2venjydoamv4ukkzofqfnc6.py 2025-12-04T09:41:13.0712851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0712944Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0713085Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0713333Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0715021Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0715104Z graph_break [] 2025-12-04T09:41:13.0715213Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0715389Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0715480Z Autotune Choices Stats: 2025-12-04T09:41:13.0716308Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_572", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0716483Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0716569Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0716683Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0717154Z triton_mm_572 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0717632Z triton_mm_573 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0718112Z triton_mm_577 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0718599Z triton_mm_580 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0719125Z triton_mm_581 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0719591Z triton_mm_570 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0720063Z triton_mm_571 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0720576Z triton_mm_574 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0721083Z triton_mm_575 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0721549Z triton_mm_576 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0721897Z SingleProcess AUTOTUNE benchmarking takes 0.2085 seconds and 0.6378 seconds precompiling for 15 choices 2025-12-04T09:41:13.0722255Z Compiled module path: /tmp/tmp5qwmyj31/f4/cf4akhntch3dmtxisx7zye35tofaic5dk6pypiqukvc5hwaop3fe.py 2025-12-04T09:41:13.0722610Z Compiled module path: /tmp/tmp5qwmyj31/ha/chawfc4mtkasifxs6qvh2eotnj3krupriktqki2yomusm5haaai6.py 2025-12-04T09:41:13.0722793Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0722888Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0723026Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0723285Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0724974Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0725065Z graph_break [] 2025-12-04T09:41:13.0725171Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0725348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0725444Z Autotune Choices Stats: 2025-12-04T09:41:13.0726462Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_603", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0726564Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0726650Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0726757Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0727287Z triton_mm_603 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0727762Z triton_mm_604 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0728231Z triton_mm_605 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0728753Z triton_mm_606 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0729231Z triton_mm_609 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0729701Z triton_mm_611 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0730176Z triton_mm_613 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0730692Z triton_mm_614 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0731163Z triton_mm_601 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0731628Z triton_mm_602 0.0287 ms 99.9% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0731969Z SingleProcess AUTOTUNE benchmarking takes 0.2008 seconds and 0.6272 seconds precompiling for 15 choices 2025-12-04T09:41:13.0732322Z Compiled module path: /tmp/tmpo9rrm3hb/j2/cj2kdgccny72hbdtda3xzix2me555tucumvq6dcuvocaexfvm3zb.py 2025-12-04T09:41:13.0732677Z Compiled module path: /tmp/tmpo9rrm3hb/4b/c4brg7cx2lwg4dsspsrd2lyeawoavk4cpqnc52xgvpv7ub4xyife.py 2025-12-04T09:41:13.0732857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0732957Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0733092Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0733349Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0735023Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0735112Z graph_break [] 2025-12-04T09:41:13.0735220Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0735724Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0735855Z Autotune Choices Stats: 2025-12-04T09:41:13.0736863Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_639", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0736965Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0737062Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0737169Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0737648Z triton_mm_639 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0738118Z triton_mm_632 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0738655Z triton_mm_638 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0739134Z triton_mm_637 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0739604Z triton_mm_633 0.0287 ms 96.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0740085Z triton_mm_643 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0740548Z triton_mm_631 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0741075Z triton_mm_634 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0741538Z triton_mm_635 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0742004Z triton_mm_636 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0742352Z SingleProcess AUTOTUNE benchmarking takes 0.2032 seconds and 0.6400 seconds precompiling for 15 choices 2025-12-04T09:41:13.0742700Z Compiled module path: /tmp/tmpvk_v1nt1/al/calvdf47t4gsdri4v57dyyzt4maptidmnfghvptx324c6ueqz473.py 2025-12-04T09:41:13.0743048Z Compiled module path: /tmp/tmpvk_v1nt1/lu/cluo2f4vnzntzynvmewbgdaorgxwvc4mjp75zmm3fsdv5nestqdm.py 2025-12-04T09:41:13.0743235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0743334Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0743473Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0743722Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0745402Z inductor [('triton_bundler_save_kernel', 248), ('benchmarking.InductorBenchmarker.benchmark_gpu', 27), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 12), ('coordesc_tuning_bench', 6), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0745490Z graph_break [] 2025-12-04T09:41:13.0745597Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0745778Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0745953Z Autotune Choices Stats: 2025-12-04T09:41:13.0746781Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_668", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0746884Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0746970Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0747078Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0747563Z triton_mm_668 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0748051Z triton_mm_673 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0748577Z triton_mm_661 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0749044Z triton_mm_662 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0749514Z triton_mm_664 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0749999Z triton_mm_666 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0750539Z triton_mm_660 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0751006Z triton_mm_663 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0751469Z triton_mm_665 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0751945Z triton_mm_669 0.0307 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0752289Z SingleProcess AUTOTUNE benchmarking takes 0.2085 seconds and 0.6332 seconds precompiling for 15 choices 2025-12-04T09:41:13.0752647Z Compiled module path: /tmp/tmp3o38vp6k/3x/c3xl4taf4zn5lox64lkqtfkt3qjcksh7v5cdlstbqy5a3odjxtz6.py 2025-12-04T09:41:13.0753005Z Compiled module path: /tmp/tmp3o38vp6k/4x/c4xjnodchj6ibvtcnowkrtx3oyfoipw5k4mysuusfoalorvm7uls.py 2025-12-04T09:41:13.0753188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0753292Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0753427Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0753682Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0755355Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0755443Z graph_break [] 2025-12-04T09:41:13.0755629Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0755806Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0755902Z Autotune Choices Stats: 2025-12-04T09:41:13.0756728Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_691", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0756822Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0756913Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0757021Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0757497Z triton_mm_691 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0758016Z triton_mm_692 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0758483Z triton_mm_693 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0758951Z triton_mm_694 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0759413Z triton_mm_695 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0759930Z triton_mm_696 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0760444Z triton_mm_697 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0760915Z triton_mm_698 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0761386Z triton_mm_699 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0761860Z triton_mm_700 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0762209Z SingleProcess AUTOTUNE benchmarking takes 0.2041 seconds and 0.6475 seconds precompiling for 15 choices 2025-12-04T09:41:13.0762568Z Compiled module path: /tmp/tmpcjdheg_x/uy/cuykhl6ore3zfznkrueimqubxt4uk3tpjenrdcqvrtamvvwmfngl.py 2025-12-04T09:41:13.0762920Z Compiled module path: /tmp/tmpcjdheg_x/aq/caqvsnyf5h5bv3f4uhfwimqvm7hf3b6we6pwdlf64itlenqopdlb.py 2025-12-04T09:41:13.0763096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0763189Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0763326Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0763576Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0765334Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0765426Z graph_break [] 2025-12-04T09:41:13.0765531Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0765710Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0765802Z Autotune Choices Stats: 2025-12-04T09:41:13.0766625Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_725", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0766719Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0766803Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0766917Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0767482Z triton_mm_725 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0767959Z triton_mm_731 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0768421Z triton_mm_723 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0768886Z triton_mm_726 0.0287 ms 99.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0769358Z triton_mm_728 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0769903Z triton_mm_729 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0770375Z triton_mm_730 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0770848Z triton_mm_732 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0771325Z triton_mm_733 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0771798Z triton_mm_734 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0772148Z SingleProcess AUTOTUNE benchmarking takes 0.2070 seconds and 0.6359 seconds precompiling for 15 choices 2025-12-04T09:41:13.0772505Z Compiled module path: /tmp/tmpcri9z8ii/ll/clleccci72holl57y6qdvpjw4nqe2v3o33e3b4dbi52l4yonc2u4.py 2025-12-04T09:41:13.0772863Z Compiled module path: /tmp/tmpcri9z8ii/nt/cntotzvaz6lwon2hethxzh4kfya2upvsk7elnoqjviiakmz27h65.py 2025-12-04T09:41:13.0773046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0773140Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0773275Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0773529Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0775283Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0775381Z graph_break [] 2025-12-04T09:41:13.0775486Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0775663Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0775756Z Autotune Choices Stats: 2025-12-04T09:41:13.0776581Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_758", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8", "best_time": 0.027744000777602196, "best_triton_pos": 0} 2025-12-04T09:41:13.0776679Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0776809Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0776917Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0777402Z triton_mm_758 0.0277 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0777873Z triton_mm_753 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0778339Z triton_mm_755 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0778811Z triton_mm_759 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0779323Z triton_mm_760 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0779802Z triton_mm_761 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0780279Z triton_mm_764 0.0287 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0780745Z triton_mm_752 0.0287 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0781210Z triton_mm_751 0.0297 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0781688Z triton_mm_754 0.0297 ms 93.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0782028Z SingleProcess AUTOTUNE benchmarking takes 0.2023 seconds and 0.6256 seconds precompiling for 15 choices 2025-12-04T09:41:13.0782368Z Compiled module path: /tmp/tmp_crk5itx/7h/c7hlonnyw5ynm5th3n7w5hspkcrh4z7yqrc5z26yajh3wiscnl4h.py 2025-12-04T09:41:13.0782725Z Compiled module path: /tmp/tmp_crk5itx/er/cerbi2equovlvhddb3tgbmag3wpo2vfegtrxofu5awfig5xspie5.py 2025-12-04T09:41:13.0782902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0782998Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0783133Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0783383Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0785145Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0785231Z graph_break [] 2025-12-04T09:41:13.0785340Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0785516Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0785609Z Autotune Choices Stats: 2025-12-04T09:41:13.0786434Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_781", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.02969600073993206, "best_triton_pos": 0} 2025-12-04T09:41:13.0786573Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0786666Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0786773Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0787250Z triton_mm_781 0.0297 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0787731Z triton_mm_794 0.0297 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0788201Z triton_mm_780 0.0307 ms 96.8% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0788712Z triton_mm_787 0.0307 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0789192Z triton_mm_789 0.0307 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0789663Z triton_mm_790 0.0307 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0790192Z triton_mm_792 0.0307 ms 96.7% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0790661Z triton_mm_788 0.0308 ms 96.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0791133Z triton_mm_782 0.0317 ms 93.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0791607Z triton_mm_783 0.0317 ms 93.5% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0791953Z SingleProcess AUTOTUNE benchmarking takes 0.2089 seconds and 0.6275 seconds precompiling for 15 choices 2025-12-04T09:41:13.0792312Z Compiled module path: /tmp/tmpat625dol/5j/c5jcehve6psqmluc2fkhixgt3hzwttzv5cbgxyzh3di3c76stb5j.py 2025-12-04T09:41:13.0792667Z Compiled module path: /tmp/tmpat625dol/kn/cknktcrdvm4elzlwkhq5s63lc7s4ucgqeicjmy6tacyszdhzz5a3.py 2025-12-04T09:41:13.0792850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0792946Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0793081Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0793341Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0795091Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0795181Z graph_break [] 2025-12-04T09:41:13.0795287Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0795467Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0795559Z Autotune Choices Stats: 2025-12-04T09:41:13.0796386Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_810", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0796524Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0796609Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0796717Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0797195Z triton_mm_810 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0797663Z triton_mm_812 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0798138Z triton_mm_813 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0798654Z triton_mm_815 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0799123Z triton_mm_816 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0799599Z triton_mm_817 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0800070Z triton_mm_819 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0800551Z triton_mm_820 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0801037Z triton_mm_821 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0801519Z triton_mm_811 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0801861Z SingleProcess AUTOTUNE benchmarking takes 0.2007 seconds and 0.6373 seconds precompiling for 15 choices 2025-12-04T09:41:13.0802222Z Compiled module path: /tmp/tmp1pvy590j/wz/cwz3w7hwh6gvj22bftrdtw7bfbogxwogh7rm4vahl7d4y2ddh2u5.py 2025-12-04T09:41:13.0802577Z Compiled module path: /tmp/tmp1pvy590j/y2/cy2v3dpgqco7sv23sigs3qjjosno5eomtrgfwlewtcvvuagtado3.py 2025-12-04T09:41:13.0802755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0802862Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0803005Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0803337Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0805022Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0805108Z graph_break [] 2025-12-04T09:41:13.0805221Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0805402Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0805570Z Autotune Choices Stats: 2025-12-04T09:41:13.0806397Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_846", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0806492Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0806587Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0806701Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0807221Z triton_mm_846 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0807700Z triton_mm_847 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0808221Z triton_mm_849 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0808715Z triton_mm_853 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0809185Z triton_mm_848 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0809660Z triton_mm_850 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0810137Z triton_mm_851 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0810626Z triton_mm_852 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=8 2025-12-04T09:41:13.0811102Z triton_mm_854 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0811567Z triton_mm_845 0.0308 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0811925Z SingleProcess AUTOTUNE benchmarking takes 0.2058 seconds and 0.6435 seconds precompiling for 15 choices 2025-12-04T09:41:13.0812286Z Compiled module path: /tmp/tmpwhwf0ibm/7v/c7v6bgeyvxnrvpshs2r4f3wzf5esyopbsdidn2nmcac4x5wjjwnx.py 2025-12-04T09:41:13.0812658Z Compiled module path: /tmp/tmpwhwf0ibm/tf/ctf7xbdusnrfspyrmou3lxcbkogwc2fspbbmznenz7ifzfbfs456.py 2025-12-04T09:41:13.0812844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0812944Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0813166Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0813420Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0815103Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0815193Z graph_break [] 2025-12-04T09:41:13.0815300Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0815531Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0819339Z Autotune Choices Stats: 2025-12-04T09:41:13.0820247Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_871", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0820343Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0820430Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0820544Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0821021Z triton_mm_871 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0821554Z triton_mm_872 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0822029Z triton_mm_873 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0822491Z triton_mm_874 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0822960Z triton_mm_875 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0823422Z triton_mm_876 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0823900Z triton_mm_877 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0824378Z triton_mm_878 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0824847Z triton_mm_879 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0825321Z triton_mm_880 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0825661Z SingleProcess AUTOTUNE benchmarking takes 0.2002 seconds and 0.6285 seconds precompiling for 15 choices 2025-12-04T09:41:13.0826030Z Compiled module path: /tmp/tmpmwjjx73y/wb/cwbsd56cyejfu6yitythbmga3wllzq4qb3mhdxxhxv7igpsk6rkf.py 2025-12-04T09:41:13.0826388Z Compiled module path: /tmp/tmpmwjjx73y/xu/cxuzsholowpf77keyfoipo3tmz3nako2i2tsxuhkjqzbk2unl3bz.py 2025-12-04T09:41:13.0826650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0826754Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0826889Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0827145Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0828831Z inductor [('triton_bundler_save_kernel', 224), ('benchmarking.InductorBenchmarker.benchmark_gpu', 24), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 9), ('coordesc_tuning_bench', 3), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0828962Z graph_break [] 2025-12-04T09:41:13.0829066Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0829257Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0829351Z Autotune Choices Stats: 2025-12-04T09:41:13.0830173Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_902", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0830269Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0830356Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0830467Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0830952Z triton_mm_902 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0831472Z triton_mm_903 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0831943Z triton_mm_905 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0832429Z triton_mm_907 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0832902Z triton_mm_908 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0833382Z triton_mm_909 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0833864Z triton_mm_911 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0834345Z triton_mm_913 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0834807Z triton_mm_900 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0835622Z triton_mm_901 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0836059Z SingleProcess AUTOTUNE benchmarking takes 0.2012 seconds and 0.7632 seconds precompiling for 15 choices 2025-12-04T09:41:13.0836416Z Compiled module path: /tmp/tmphyk566_q/yu/cyuea4r4os7pugqxnksie3dv6pq7fo2ns77dqcfthvdvgf4r52j3.py 2025-12-04T09:41:13.0836918Z Compiled module path: /tmp/tmphyk566_q/ly/clymrvghrylu2ct5p7svacigybspd2ksvswnqbpvhrgbdyv7v3tn.py 2025-12-04T09:41:13.0837101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0837196Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0837342Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0837596Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0839292Z inductor [('triton_bundler_save_kernel', 232), ('benchmarking.InductorBenchmarker.benchmark_gpu', 25), ('async_compile_cache_miss', 23), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('benchmarking.InductorBenchmarker.benchmark', 10), ('coordesc_tuning_bench', 4), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0839429Z graph_break [] 2025-12-04T09:41:13.0839546Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0839722Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0839814Z Autotune Choices Stats: 2025-12-04T09:41:13.0840643Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_939", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0840737Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0840822Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0840992Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0841474Z triton_mm_939 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0841943Z triton_mm_931 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0842405Z triton_mm_932 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0842874Z triton_mm_933 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0843339Z triton_mm_935 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0843814Z triton_mm_937 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0844284Z triton_mm_941 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0844750Z triton_mm_940 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0845222Z triton_mm_944 0.0296 ms 93.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0845685Z triton_mm_938 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0846033Z SingleProcess AUTOTUNE benchmarking takes 0.2030 seconds and 0.6686 seconds precompiling for 15 choices 2025-12-04T09:41:13.0846504Z Compiled module path: /tmp/tmp4cvs758s/n6/cn6tmzlh3kkhsuhuer55fd5jb2spgvj2yyqbmx3rllxtqyhuj3t2.py 2025-12-04T09:41:13.0846854Z Compiled module path: /tmp/tmp4cvs758s/6r/c6r375atdd4kvestjgimuafqdjuwpz5lxwv7wbi4jwlyczo6vzzt.py 2025-12-04T09:41:13.0847093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0847188Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0847328Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0847611Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0849800Z inductor [('triton_bundler_save_kernel', 288), ('benchmarking.InductorBenchmarker.benchmark_gpu', 37), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 22), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('coordesc_tuning_bench', 15), ('generated_module_cache_hit', 15), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0849938Z graph_break [] 2025-12-04T09:41:13.0850047Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0850223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0850317Z Autotune Choices Stats: 2025-12-04T09:41:13.0851146Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_974", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.027648000046610832, "best_triton_pos": 0} 2025-12-04T09:41:13.0851287Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0851373Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0851480Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0851975Z triton_mm_974 0.0276 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0852443Z triton_mm_961 0.0287 ms 96.4% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0852917Z triton_mm_969 0.0287 ms 96.3% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0853383Z triton_mm_967 0.0297 ms 93.2% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0853862Z triton_mm_962 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0854322Z triton_mm_963 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0854786Z triton_mm_965 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0855252Z triton_mm_966 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0855716Z triton_mm_968 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0856270Z triton_mm_970 0.0297 ms 93.1% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=128, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0856614Z SingleProcess AUTOTUNE benchmarking takes 0.2037 seconds and 0.6293 seconds precompiling for 15 choices 2025-12-04T09:41:13.0856976Z Compiled module path: /tmp/tmpzrw3psn5/ex/cexdxarjh6qlkmrh5ex2jzrhiw737frfohgvjrzxgquhfv5g2o4l.py 2025-12-04T09:41:13.0857336Z Compiled module path: /tmp/tmpzrw3psn5/3v/c3vwhsbwrmtggganhnbbkwvtdwpm4qfdu33rqhqhbdczjvs5z6rs.py 2025-12-04T09:41:13.0857513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:41:13.0857613Z frames [('total', 2), ('ok', 2)] 2025-12-04T09:41:13.0857747Z stats [('calls_captured', 4), ('unique_graphs', 2)] 2025-12-04T09:41:13.0857996Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:41:13.0859777Z inductor [('triton_bundler_save_kernel', 280), ('benchmarking.InductorBenchmarker.benchmark_gpu', 35), ('async_compile_cache_miss', 31), ('benchmarking.InductorBenchmarker.benchmark', 20), ('generated_module_cache_miss', 15), ('select_algorithm_num_precompiles', 15), ('generated_module_cache_hit', 15), ('coordesc_tuning_bench', 13), ('fxgraph_cache_miss', 2), ('pattern_matcher_count', 2), ('pattern_matcher_nodes', 2), ('async_compile_cache_hit', 2), ('triton_bundler_save_static_autotuner', 2), ('select_algorithm_precompile', 1), ('select_algorithm_autotune', 1)] 2025-12-04T09:41:13.0859902Z graph_break [] 2025-12-04T09:41:13.0860010Z aten_mm_info [('aten.mm_256_256_256', 2)] 2025-12-04T09:41:13.0860186Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:41:13.0860280Z Autotune Choices Stats: 2025-12-04T09:41:13.0861117Z {"num_choices": 15, "num_triton_choices": 15, "best_kernel": "triton_mm_1004", "best_kernel_desc": "ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4", "best_time": 0.028672000393271446, "best_triton_pos": 0} 2025-12-04T09:41:13.0861261Z AUTOTUNE mm(256x256, 256x256) 2025-12-04T09:41:13.0861351Z strides: [256, 1], [1, 256] 2025-12-04T09:41:13.0861458Z dtypes: torch.float16, torch.float16 2025-12-04T09:41:13.0861940Z triton_mm_1004 0.0287 ms 100.0% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0862405Z triton_mm_990 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=16, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=1, num_warps=2 2025-12-04T09:41:13.0862875Z triton_mm_991 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=32, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0863353Z triton_mm_992 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=32, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0863822Z triton_mm_993 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=32, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=5, num_warps=8 2025-12-04T09:41:13.0864289Z triton_mm_995 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=2, num_warps=4 2025-12-04T09:41:13.0864750Z triton_mm_996 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=8 2025-12-04T09:41:13.0865222Z triton_mm_997 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0865694Z triton_mm_998 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=32, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=4, num_warps=8 2025-12-04T09:41:13.0866247Z triton_mm_999 0.0297 ms 96.6% ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=64, BLOCK_N=128, EVEN_K=True, GROUP_M=8, USE_FAST_ACCUM=False, num_stages=3, num_warps=4 2025-12-04T09:41:13.0866593Z SingleProcess AUTOTUNE benchmarking takes 0.2117 seconds and 0.6653 seconds precompiling for 15 choices 2025-12-04T09:41:13.0866954Z Compiled module path: /tmp/tmpqepf3qkx/w7/cw7mt2dfsg6awmx5ypnrdsmu6tzcfbd2kajbvpekvwnrvbl2ehhm.py 2025-12-04T09:41:13.0867302Z Compiled module path: /tmp/tmpqepf3qkx/n7/cn7cj66zjlnxk7nmy3mbawg3llytcb7qjbq6w6mqi633xq673ums.py 2025-12-04T09:41:13.0867931Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_benchmark_fusion/inductor.test_benchmark_fusion-1a5c9a2f1b3021da.xml - 2025-12-04T09:41:13.0868081Z =========================== short test summary info ============================ 2025-12-04T09:41:13.0868877Z FAILED [4.3923s] inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code - RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0868961Z Searched string: 2025-12-04T09:41:13.0869069Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0869171Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0869344Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0869591Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0869691Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0869905Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0870008Z del arg0_1 2025-12-04T09:41:13.0870147Z del arg2_1 2025-12-04T09:41:13.0870269Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0870544Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0870649Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0870852Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0870932Z del arg1_1 2025-12-04T09:41:13.0871017Z return (buf1, ) 2025-12-04T09:41:13.0871022Z 2025-12-04T09:41:13.0871128Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0871211Z call = runner.call 2025-12-04T09:41:13.0871373Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0871378Z 2025-12-04T09:41:13.0871382Z 2025-12-04T09:41:13.0871529Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0871661Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0871814Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0872027Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0872221Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0872428Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0872544Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0872712Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0872717Z 2025-12-04T09:41:13.0872724Z 2025-12-04T09:41:13.0872815Z if __name__ == "__main__": 2025-12-04T09:41:13.0873020Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0873187Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0873305Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0873309Z 2025-12-04T09:41:13.0873313Z 2025-12-04T09:41:13.0873493Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0874163Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0874256Z 2025-12-04T09:41:13.0874478Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0875210Z FAILED [4.3588s] inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code - RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0875293Z Searched string: 2025-12-04T09:41:13.0875401Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0875502Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0875665Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0875906Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0876012Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0876266Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0876349Z del arg0_1 2025-12-04T09:41:13.0876438Z del arg2_1 2025-12-04T09:41:13.0876537Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0876805Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0876902Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0877104Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0877190Z del arg1_1 2025-12-04T09:41:13.0877277Z return (buf1, ) 2025-12-04T09:41:13.0877281Z 2025-12-04T09:41:13.0877386Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0877466Z call = runner.call 2025-12-04T09:41:13.0877625Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0877672Z 2025-12-04T09:41:13.0877675Z 2025-12-04T09:41:13.0877826Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0877963Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0878123Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0878326Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0878520Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0878726Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0878838Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0879004Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0879009Z 2025-12-04T09:41:13.0879013Z 2025-12-04T09:41:13.0879102Z if __name__ == "__main__": 2025-12-04T09:41:13.0879306Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0879474Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0879595Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0879600Z 2025-12-04T09:41:13.0879603Z 2025-12-04T09:41:13.0879795Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0880505Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0880511Z 2025-12-04T09:41:13.0880730Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0881459Z FAILED [3.9186s] inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code - RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0881544Z Searched string: 2025-12-04T09:41:13.0881652Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0881758Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0881923Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0882357Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0882458Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0882670Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0882754Z del arg0_1 2025-12-04T09:41:13.0882833Z del arg2_1 2025-12-04T09:41:13.0882932Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0883201Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0883297Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0883496Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0883578Z del arg1_1 2025-12-04T09:41:13.0883663Z return (buf1, ) 2025-12-04T09:41:13.0883667Z 2025-12-04T09:41:13.0883813Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0883896Z call = runner.call 2025-12-04T09:41:13.0884058Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0884067Z 2025-12-04T09:41:13.0884071Z 2025-12-04T09:41:13.0884219Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0884355Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0884503Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0884714Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0884904Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0885110Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0885225Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0885394Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0885446Z 2025-12-04T09:41:13.0885452Z 2025-12-04T09:41:13.0885544Z if __name__ == "__main__": 2025-12-04T09:41:13.0885750Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0885920Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0886039Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0886044Z 2025-12-04T09:41:13.0886047Z 2025-12-04T09:41:13.0886226Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0886877Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0886882Z 2025-12-04T09:41:13.0887150Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0887880Z FAILED [3.6587s] inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code - RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0887970Z Searched string: 2025-12-04T09:41:13.0888077Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0888184Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0888351Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0888591Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0888692Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0888903Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0888985Z del arg0_1 2025-12-04T09:41:13.0889063Z del arg2_1 2025-12-04T09:41:13.0889161Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0889427Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0889528Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0889729Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0889809Z del arg1_1 2025-12-04T09:41:13.0889983Z return (buf1, ) 2025-12-04T09:41:13.0889988Z 2025-12-04T09:41:13.0890091Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0890173Z call = runner.call 2025-12-04T09:41:13.0890332Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0890336Z 2025-12-04T09:41:13.0890340Z 2025-12-04T09:41:13.0890488Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0890620Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0890767Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0890975Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0891166Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0891369Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0891521Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0891697Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0891701Z 2025-12-04T09:41:13.0891705Z 2025-12-04T09:41:13.0891797Z if __name__ == "__main__": 2025-12-04T09:41:13.0891998Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0892162Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0892280Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0892285Z 2025-12-04T09:41:13.0892288Z 2025-12-04T09:41:13.0892466Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0893122Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0893166Z 2025-12-04T09:41:13.0893385Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0894124Z FAILED [4.0362s] inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code - RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0894213Z Searched string: 2025-12-04T09:41:13.0894319Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0894423Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0894587Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0894822Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0894923Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0895132Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0895216Z del arg0_1 2025-12-04T09:41:13.0895297Z del arg2_1 2025-12-04T09:41:13.0895398Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0895665Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0895766Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0895966Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0896049Z del arg1_1 2025-12-04T09:41:13.0896131Z return (buf1, ) 2025-12-04T09:41:13.0896135Z 2025-12-04T09:41:13.0896237Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0896319Z call = runner.call 2025-12-04T09:41:13.0896477Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0896481Z 2025-12-04T09:41:13.0896485Z 2025-12-04T09:41:13.0896633Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0896763Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0896914Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0897122Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0897316Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0897595Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0897709Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0897875Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0897880Z 2025-12-04T09:41:13.0897883Z 2025-12-04T09:41:13.0897974Z if __name__ == "__main__": 2025-12-04T09:41:13.0898176Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0898342Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0898462Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0898466Z 2025-12-04T09:41:13.0898470Z 2025-12-04T09:41:13.0898649Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0899344Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0899348Z 2025-12-04T09:41:13.0899568Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0900348Z FAILED [4.2237s] inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code - RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0900432Z Searched string: 2025-12-04T09:41:13.0900538Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0900642Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0900807Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0901046Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0901191Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0901405Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0901487Z del arg0_1 2025-12-04T09:41:13.0901574Z del arg2_1 2025-12-04T09:41:13.0901673Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0901938Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0902036Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0902236Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0902318Z del arg1_1 2025-12-04T09:41:13.0902400Z return (buf1, ) 2025-12-04T09:41:13.0902404Z 2025-12-04T09:41:13.0902505Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0902587Z call = runner.call 2025-12-04T09:41:13.0902747Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0902754Z 2025-12-04T09:41:13.0902758Z 2025-12-04T09:41:13.0902909Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0903042Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0903196Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0903409Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0903596Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0903795Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0903905Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0904073Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0904078Z 2025-12-04T09:41:13.0904081Z 2025-12-04T09:41:13.0904176Z if __name__ == "__main__": 2025-12-04T09:41:13.0904378Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0904545Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0904666Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0904670Z 2025-12-04T09:41:13.0904674Z 2025-12-04T09:41:13.0904938Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0905595Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0905599Z 2025-12-04T09:41:13.0905819Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0906545Z FAILED [4.2954s] inductor/test_benchmark_fusion.py::BenchmarkMultiTemplateFusionGpuTest::test_equivalent_template_code - RuntimeError: Expected to find "triton_tem_fused_addmm_relu_t_0" but did not find it 2025-12-04T09:41:13.0906628Z Searched string: 2025-12-04T09:41:13.0906734Z with torch.cuda._DeviceGuard(0): 2025-12-04T09:41:13.0906843Z torch.cuda.set_device(0) 2025-12-04T09:41:13.0907050Z buf0 = empty_strided_cuda((256, 256), (256, 1), torch.float16) 2025-12-04T09:41:13.0907290Z # Topologically Sorted Source Nodes: [a], Original ATen: [aten.t, aten.addmm] 2025-12-04T09:41:13.0907389Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0907600Z triton_tem_fused_addmm_t_0.run(arg2_1, arg0_1, buf0, 4, 1, 1, stream=stream0) 2025-12-04T09:41:13.0907683Z del arg0_1 2025-12-04T09:41:13.0907762Z del arg2_1 2025-12-04T09:41:13.0907863Z buf1 = buf0; del buf0 # reuse 2025-12-04T09:41:13.0908125Z # Topologically Sorted Source Nodes: [a, relu], Original ATen: [aten.addmm, aten.relu] 2025-12-04T09:41:13.0908227Z stream0 = get_raw_stream(0) 2025-12-04T09:41:13.0908426Z triton_poi_fused_addmm_relu_1.run(buf1, arg1_1, 65536, stream=stream0) 2025-12-04T09:41:13.0908508Z del arg1_1 2025-12-04T09:41:13.0908663Z return (buf1, ) 2025-12-04T09:41:13.0908667Z 2025-12-04T09:41:13.0908774Z runner = Runner(partitions=[]) 2025-12-04T09:41:13.0908854Z call = runner.call 2025-12-04T09:41:13.0909012Z recursively_apply_fns = runner.recursively_apply_fns 2025-12-04T09:41:13.0909021Z 2025-12-04T09:41:13.0909025Z 2025-12-04T09:41:13.0909167Z def benchmark_compiled_module(times=10, repeat=10): 2025-12-04T09:41:13.0909299Z from torch._dynamo.testing import rand_strided 2025-12-04T09:41:13.0909447Z from torch._inductor.utils import print_performance 2025-12-04T09:41:13.0909658Z arg0_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0909850Z arg1_1 = rand_strided((256, ), (1, ), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0910058Z arg2_1 = rand_strided((256, 256), (256, 1), device='cuda:0', dtype=torch.float16) 2025-12-04T09:41:13.0910188Z fn = lambda: call([arg0_1, arg1_1, arg2_1]) 2025-12-04T09:41:13.0910378Z return print_performance(fn, times=times, repeat=repeat) 2025-12-04T09:41:13.0910387Z 2025-12-04T09:41:13.0910393Z 2025-12-04T09:41:13.0910483Z if __name__ == "__main__": 2025-12-04T09:41:13.0910685Z from torch._inductor.wrapper_benchmark import compiled_module_main 2025-12-04T09:41:13.0910852Z compiled_module_main('None', benchmark_compiled_module) 2025-12-04T09:41:13.0910973Z From CHECK: triton_tem_fused_addmm_relu_t_0 2025-12-04T09:41:13.0910978Z 2025-12-04T09:41:13.0910981Z 2025-12-04T09:41:13.0911160Z To execute this test, run the following from the base repo dir: 2025-12-04T09:41:13.0911812Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_benchmark_fusion.py BenchmarkMultiTemplateFusionGpuTest.test_equivalent_template_code 2025-12-04T09:41:13.0911817Z 2025-12-04T09:41:13.0912036Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:41:13.0912209Z ============= 7 failed, 43 passed, 50 skipped in 140.08s (0:02:20) ============= 2025-12-04T09:41:13.0912216Z 2025-12-04T09:41:13.0912696Z FINISHED PRINTING LOG FILE of inductor/test_benchmark_fusion 1/1 (test/test-reports/inductor.test_benchmark_fusion_1.1_baf135911d2d356d_.log) 2025-12-04T09:41:13.0912703Z 2025-12-04T09:41:13.0913087Z Finished inductor/test_benchmark_fusion 1/1 ... [2025-12-04 09:41:12.888472][1775.304938873], took 2.47min 2025-12-04T09:41:13.0913753Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_benchmark_fusion/inductor.test_benchmark_fusion-1a5c9a2f1b3021da.xml 2025-12-04T09:41:13.4615814Z Uploading logs for 57118183167 to S3 2025-12-04T09:41:13.5055430Z Uploading artifacts took 0.42 seconds 2025-12-04T09:41:13.5055880Z inductor/test_benchmark_fusion 1/1 failed! 2025-12-04T09:41:13.5059476Z Running inductor/test_remote_cache 1/1 ... [2025-12-04 09:41:13.505676][1775.92214798] 2025-12-04T09:41:13.5060088Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:41:13.5063993Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_remote_cache.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:41:13.506043] 2025-12-04T09:41:16.7770857Z 2025-12-04T09:41:16.7771875Z inductor/test_remote_cache 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_remote_cache_1.1_c953161b101bcc90_.log 2025-12-04T09:41:16.7772665Z Running 0 items in this shard: 2025-12-04T09:41:16.7772856Z 2025-12-04T09:41:16.7773133Z Finished inductor/test_remote_cache 1/1 ... [2025-12-04 09:41:16.776748][1779.193218672], took 0.05min 2025-12-04T09:41:16.7826036Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_remote_cache/inductor.test_remote_cache-e53c2a7e4fbe3b3f.xml 2025-12-04T09:41:16.8096397Z Running inductor/test_coordinate_descent_tuner 1/1 ... [2025-12-04 09:41:16.809312][1779.22578637] 2025-12-04T09:41:16.8097414Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:41:16.8100335Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_coordinate_descent_tuner.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:41:16.809612] 2025-12-04T09:41:22.9857985Z 2025-12-04T09:41:22.9859165Z inductor/test_coordinate_descent_tuner 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_coordinate_descent_tuner_1.1_d79d4fca9c00c8c8_.log 2025-12-04T09:41:22.9860025Z Running 0 items in this shard: 2025-12-04T09:41:22.9860214Z 2025-12-04T09:41:22.9860545Z Finished inductor/test_coordinate_descent_tuner 1/1 ... [2025-12-04 09:41:22.985455][1785.401927772], took 0.10min 2025-12-04T09:41:22.9914592Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_coordinate_descent_tuner/inductor.test_coordinate_descent_tuner-9ed05914eb073327.xml 2025-12-04T09:41:23.0646261Z Running inductor/test_inplace_padding 1/1 ... [2025-12-04 09:41:23.064274][1785.480746886] 2025-12-04T09:41:23.0646900Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:41:23.0650323Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_inplace_padding.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:41:23.064621] 2025-12-04T09:45:14.1122028Z 2025-12-04T09:45:14.1123283Z inductor/test_inplace_padding 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_inplace_padding_1.1_85df0d7c6ccefed5_.log 2025-12-04T09:45:14.1144061Z Running 50 items in this shard: test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune 2025-12-04T09:45:14.1161628Z 2025-12-04T09:45:14.1161937Z Finished inductor/test_inplace_padding 1/1 ... [2025-12-04 09:45:14.111634][2016.528107362], took 3.85min 2025-12-04T09:45:14.1180369Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_inplace_padding/inductor.test_inplace_padding-ec8a12f82b4c1954.xml 2025-12-04T09:45:14.2093810Z Running inductor/test_cudacodecache 1/1 ... [2025-12-04 09:45:14.208503][2016.624972826] 2025-12-04T09:45:14.2094364Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:45:14.2095611Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cudacodecache.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:14.208864] 2025-12-04T09:45:22.1384017Z 2025-12-04T09:45:22.1384829Z PRINTING LOG FILE of inductor/test_cudacodecache 1/1 (test/test-reports/inductor.test_cudacodecache_1.1_a3b71ead0c191355_.log) 2025-12-04T09:45:22.1386028Z Test results will be stored in test-reports/python-pytest/inductor.test_cudacodecache/inductor.test_cudacodecache-c55849964bd60b32.xml 2025-12-04T09:45:22.1386859Z ============================= test session starts ============================== 2025-12-04T09:45:22.1387422Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:45:22.1388015Z cachedir: .pytest_cache 2025-12-04T09:45:22.1388846Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:45:22.1389476Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:45:22.1389778Z configfile: pytest.ini 2025-12-04T09:45:22.1390365Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:45:22.1391034Z collecting ... collected 3 items 2025-12-04T09:45:22.1391386Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:45:22.1405364Z Running 50 items in this shard: test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load, test/inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load 2025-12-04T09:45:22.1419289Z 2025-12-04T09:45:22.1419591Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load PASSED [1.2006s] [ 2%] 2025-12-04T09:45:22.1420629Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.798000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpte0fa52g 2025-12-04T09:45:22.1421443Z FAILED [0.0290s] [ 2%] 2025-12-04T09:45:22.1422229Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.829000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpg87i14ht 2025-12-04T09:45:22.1423046Z FAILED [0.0051s] [ 2%] 2025-12-04T09:45:22.1423827Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.835000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpiocl98_k 2025-12-04T09:45:22.1424638Z FAILED [0.0050s] [ 2%] 2025-12-04T09:45:22.1425423Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.841000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpwocaqjac 2025-12-04T09:45:22.1426237Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1427016Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.847000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmptu78519d 2025-12-04T09:45:22.1427867Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1428643Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.853000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpnuyxk6tp 2025-12-04T09:45:22.1429453Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1430239Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.862000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp5erqqg9r 2025-12-04T09:45:22.1431047Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1431917Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.868000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpdx1u0n5u 2025-12-04T09:45:22.1432732Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1433503Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.874000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp4ccd98gx 2025-12-04T09:45:22.1442001Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1442839Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.880000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpzsxri6pe 2025-12-04T09:45:22.1443661Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1444449Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.885000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpjvimccfp 2025-12-04T09:45:22.1445412Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1446178Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.891000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmptjyw2wai 2025-12-04T09:45:22.1446988Z FAILED [0.0047s] [ 2%] 2025-12-04T09:45:22.1447824Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.897000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpuu2nrrsj 2025-12-04T09:45:22.1448629Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1449401Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.905000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpatij2hy5 2025-12-04T09:45:22.1450283Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1451065Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.911000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmplgbk7opj 2025-12-04T09:45:22.1451873Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1452644Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.917000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp4andu5c9 2025-12-04T09:45:22.1453449Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1454222Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.923000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp3vdisflw 2025-12-04T09:45:22.1455017Z FAILED [0.0048s] [ 2%] 2025-12-04T09:45:22.1455796Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.929000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpeq2rx1l9 2025-12-04T09:45:22.1456607Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1457590Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.934000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmphhy708fb 2025-12-04T09:45:22.1458526Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1459318Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.940000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpclsrypx0 2025-12-04T09:45:22.1460129Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1460902Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.948000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpnxcmwyxa 2025-12-04T09:45:22.1461699Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1462474Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.954000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpufqge822 2025-12-04T09:45:22.1463289Z FAILED [0.0047s] [ 2%] 2025-12-04T09:45:22.1464202Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.960000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp6xmexm2n 2025-12-04T09:45:22.1465005Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1465776Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.966000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpr4td54fp 2025-12-04T09:45:22.1466582Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1467354Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.971000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmphpr76a80 2025-12-04T09:45:22.1468152Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1468923Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.977000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpoa1lvvd6 2025-12-04T09:45:22.1469775Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1470547Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.983000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp9ay9g3yr 2025-12-04T09:45:22.1471340Z FAILED [0.0048s] [ 2%] 2025-12-04T09:45:22.1472106Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.991000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpgkkd5drg 2025-12-04T09:45:22.1472907Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1473677Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:20.997000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp3rsi1l0p 2025-12-04T09:45:22.1474527Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1475301Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.003000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp22f45f3e 2025-12-04T09:45:22.1476103Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1476874Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.008000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpt7mxmufj 2025-12-04T09:45:22.1477671Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1478437Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.014000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpkhm90xxm 2025-12-04T09:45:22.1479239Z FAILED [0.0048s] [ 2%] 2025-12-04T09:45:22.1479997Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.020000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp6w8mxtue 2025-12-04T09:45:22.1480804Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1481583Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.026000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpn6o4imml 2025-12-04T09:45:22.1482387Z FAILED [0.0049s] [ 2%] 2025-12-04T09:45:22.1483153Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.035000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp2c6ectew 2025-12-04T09:45:22.1483965Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1484732Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.040000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp3ygb2vmg 2025-12-04T09:45:22.1485537Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1486300Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.046000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp4rwvtqht 2025-12-04T09:45:22.1487188Z FAILED [0.0048s] [ 2%] 2025-12-04T09:45:22.1488041Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.052000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpk3e_nrra 2025-12-04T09:45:22.1488844Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1489610Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.058000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpzjtah27_ 2025-12-04T09:45:22.1490420Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1491193Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.063000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmptr9ra1ie 2025-12-04T09:45:22.1491998Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1492815Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.069000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpuhji2gnx 2025-12-04T09:45:22.1493621Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1494394Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.077000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpu5y8xpyt 2025-12-04T09:45:22.1495201Z FAILED [0.0047s] [ 2%] 2025-12-04T09:45:22.1495971Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.083000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmppogfyzc5 2025-12-04T09:45:22.1496780Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1497613Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.089000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpyof3f3ll 2025-12-04T09:45:22.1498468Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1499237Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.095000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp560zqovg 2025-12-04T09:45:22.1500054Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1500826Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.100000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp0d47dy2a 2025-12-04T09:45:22.1501631Z FAILED [0.0044s] [ 2%] 2025-12-04T09:45:22.1502396Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.106000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp7xno_grn 2025-12-04T09:45:22.1503203Z FAILED [0.0047s] [ 2%] 2025-12-04T09:45:22.1503983Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.112000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmpk2n_t0n3 2025-12-04T09:45:22.1504797Z FAILED [0.0045s] [ 2%] 2025-12-04T09:45:22.1505565Z inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load W1204 09:45:21.120000 29769 site-packages/torch/_inductor/utils.py:1361] on error, temporary cache dir kept at /tmp/tmp52thnf2j 2025-12-04T09:45:22.1506372Z FAILED [0.0046s] [ 2%] 2025-12-04T09:45:22.1506527Z 2025-12-04T09:45:22.1506662Z =================================== FAILURES =================================== 2025-12-04T09:45:22.1507093Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1507507Z Traceback (most recent call last): 2025-12-04T09:45:22.1508049Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1508624Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1509222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1509807Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1510580Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1511356Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1511769Z Traceback (most recent call last): 2025-12-04T09:45:22.1512300Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1512884Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1513469Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1514039Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1514717Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1515543Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1515952Z Traceback (most recent call last): 2025-12-04T09:45:22.1516488Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1517059Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1517643Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1518216Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1518891Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1519675Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1520085Z Traceback (most recent call last): 2025-12-04T09:45:22.1520665Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1521242Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1521844Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1522427Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1523107Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1523889Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1524301Z Traceback (most recent call last): 2025-12-04T09:45:22.1524835Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1525402Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1525982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1526557Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1527317Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1528123Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1528533Z Traceback (most recent call last): 2025-12-04T09:45:22.1529069Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1529633Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1530229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1530795Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1531469Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1532250Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1532655Z Traceback (most recent call last): 2025-12-04T09:45:22.1533414Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1533991Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1534575Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1535132Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1536082Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1536860Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1537259Z Traceback (most recent call last): 2025-12-04T09:45:22.1537782Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1538425Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1539014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1539577Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1540245Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1541018Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1541498Z Traceback (most recent call last): 2025-12-04T09:45:22.1542018Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1542588Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1543173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1543811Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1544491Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1545261Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1545668Z Traceback (most recent call last): 2025-12-04T09:45:22.1546186Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1546756Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1547346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1547903Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1548573Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1549349Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1549764Z Traceback (most recent call last): 2025-12-04T09:45:22.1550286Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1550849Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1551443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1552004Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1552671Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1553446Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1553859Z Traceback (most recent call last): 2025-12-04T09:45:22.1554379Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1554957Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1555548Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1556229Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1556898Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1557671Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1558078Z Traceback (most recent call last): 2025-12-04T09:45:22.1558597Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1559166Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1559753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1560331Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1560999Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1561825Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1562239Z Traceback (most recent call last): 2025-12-04T09:45:22.1562760Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1563327Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1563927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1564487Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1565162Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1565934Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1566388Z Traceback (most recent call last): 2025-12-04T09:45:22.1566912Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1567554Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1568166Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1568769Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1569432Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1570203Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1570612Z Traceback (most recent call last): 2025-12-04T09:45:22.1571134Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1571697Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1572280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1572857Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1573525Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1574303Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1574713Z Traceback (most recent call last): 2025-12-04T09:45:22.1575236Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1575797Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1576390Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1576958Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1577622Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1578448Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1578944Z Traceback (most recent call last): 2025-12-04T09:45:22.1579478Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1580040Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1580627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1581191Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1581850Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1582624Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1583033Z Traceback (most recent call last): 2025-12-04T09:45:22.1583604Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1584159Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1584753Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1585318Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1585985Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1586747Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1587180Z Traceback (most recent call last): 2025-12-04T09:45:22.1587728Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1588282Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1588916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1589483Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1590157Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1590923Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1591330Z Traceback (most recent call last): 2025-12-04T09:45:22.1591854Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1592416Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1593007Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1593570Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1594240Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1595008Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1595416Z Traceback (most recent call last): 2025-12-04T09:45:22.1595945Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1596501Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1597094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1597664Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1598384Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1599146Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1599553Z Traceback (most recent call last): 2025-12-04T09:45:22.1600085Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1600651Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1601320Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1601887Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1602565Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1603335Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1603740Z Traceback (most recent call last): 2025-12-04T09:45:22.1604275Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1604838Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1605417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1606027Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1606708Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1607523Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1607928Z Traceback (most recent call last): 2025-12-04T09:45:22.1608450Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1609016Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1609595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1610160Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1610830Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1611644Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1612061Z Traceback (most recent call last): 2025-12-04T09:45:22.1612595Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1613158Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1613741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1614305Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1614976Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1615753Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1616151Z Traceback (most recent call last): 2025-12-04T09:45:22.1616677Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1617249Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1617847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1618418Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1619090Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1619865Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1620264Z Traceback (most recent call last): 2025-12-04T09:45:22.1620789Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1621350Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1621936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1622510Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1623184Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1624065Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1624467Z Traceback (most recent call last): 2025-12-04T09:45:22.1624991Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1625553Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1626131Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1626707Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1627425Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1628207Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1628653Z Traceback (most recent call last): 2025-12-04T09:45:22.1629193Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1629761Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1630355Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1630920Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1631594Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1632368Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1632771Z Traceback (most recent call last): 2025-12-04T09:45:22.1633301Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1633909Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1634496Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1635057Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1635880Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1636663Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1637064Z Traceback (most recent call last): 2025-12-04T09:45:22.1637641Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1638205Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1638790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1639347Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1640020Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1640796Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1641198Z Traceback (most recent call last): 2025-12-04T09:45:22.1641728Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1642297Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1642887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1643443Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1644108Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1644881Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1645291Z Traceback (most recent call last): 2025-12-04T09:45:22.1645811Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1646500Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1647101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1647760Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1648430Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1649211Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1649617Z Traceback (most recent call last): 2025-12-04T09:45:22.1650135Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1650703Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1651294Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1651915Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1652602Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1653376Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1653788Z Traceback (most recent call last): 2025-12-04T09:45:22.1654316Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1654884Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1655476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1656037Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1656719Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1657558Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1657975Z Traceback (most recent call last): 2025-12-04T09:45:22.1658495Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1659069Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1659659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1660230Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1660896Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1661672Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1662088Z Traceback (most recent call last): 2025-12-04T09:45:22.1662609Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1663180Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1663769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1664340Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1665010Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1665798Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1666224Z Traceback (most recent call last): 2025-12-04T09:45:22.1666747Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1667318Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1667904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1668483Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1669229Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1670011Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1670420Z Traceback (most recent call last): 2025-12-04T09:45:22.1670940Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1671503Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1672094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1672659Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1673322Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1674137Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1674545Z Traceback (most recent call last): 2025-12-04T09:45:22.1675079Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1675633Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1676217Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1676785Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1677464Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1678277Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1678690Z Traceback (most recent call last): 2025-12-04T09:45:22.1679216Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1679826Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1686280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1686867Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1687663Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1688468Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1688893Z Traceback (most recent call last): 2025-12-04T09:45:22.1689430Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1690001Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1690602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1691184Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1691868Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1692652Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1693074Z Traceback (most recent call last): 2025-12-04T09:45:22.1693614Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1694186Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1694785Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1695374Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1696043Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1696832Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1697259Z Traceback (most recent call last): 2025-12-04T09:45:22.1697915Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1698483Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1699082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1699656Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1700343Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1701114Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1701529Z Traceback (most recent call last): 2025-12-04T09:45:22.1702065Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1702639Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1703279Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1703856Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1704528Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1705297Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1705713Z Traceback (most recent call last): 2025-12-04T09:45:22.1706246Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1706811Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1707404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1708104Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1708783Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1709569Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1709985Z Traceback (most recent call last): 2025-12-04T09:45:22.1710518Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1711080Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1711683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1712267Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1712951Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1713726Z _______________________ TestCUDACodeCache.test_cuda_load _______________________ 2025-12-04T09:45:22.1714145Z Traceback (most recent call last): 2025-12-04T09:45:22.1714680Z File "/var/lib/jenkins/workspace/test/inductor/test_cudacodecache.py", line 48, in test_cuda_load 2025-12-04T09:45:22.1715252Z ) = CUDACodeCache.compile(_SOURCE_CODE, "o") 2025-12-04T09:45:22.1715846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/codecache.py", line 4185, in compile 2025-12-04T09:45:22.1716429Z with open(input_path, "a") as f: 2025-12-04T09:45:22.1717112Z FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1718341Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cudacodecache/inductor.test_cudacodecache-c55849964bd60b32.xml - 2025-12-04T09:45:22.1719210Z =========================== short test summary info ============================ 2025-12-04T09:45:22.1720277Z FAILED [0.0290s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1722043Z FAILED [0.0051s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1723709Z FAILED [0.0050s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1725360Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1727025Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1728777Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1730442Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1732104Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1733764Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1735696Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1737356Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1739020Z FAILED [0.0047s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1740677Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1742343Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1743998Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1745668Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1747391Z FAILED [0.0048s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1749182Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1750851Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1752504Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1754151Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1755870Z FAILED [0.0047s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1757528Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1759182Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1760848Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1762577Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1764229Z FAILED [0.0048s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1765887Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1767640Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1769352Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1771023Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1772667Z FAILED [0.0048s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1774320Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1775978Z FAILED [0.0049s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1777791Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1779460Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1781120Z FAILED [0.0048s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1782765Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1784480Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1786134Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1787794Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1789459Z FAILED [0.0047s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1791163Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1792820Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1794476Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1796136Z FAILED [0.0044s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1797849Z FAILED [0.0047s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1799511Z FAILED [0.0045s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1801158Z FAILED [0.0046s] inductor/test_cudacodecache.py::TestCUDACodeCache::test_cuda_load - FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpskzi7v87/ru/crur4ger33s5y4rafupzaofwdiqwss6m3i2xanfyukhomkk7y7u4.cu' 2025-12-04T09:45:22.1802182Z ========================= 49 failed, 1 passed in 1.56s ========================= 2025-12-04T09:45:22.1802452Z 2025-12-04T09:45:22.1802913Z FINISHED PRINTING LOG FILE of inductor/test_cudacodecache 1/1 (test/test-reports/inductor.test_cudacodecache_1.1_a3b71ead0c191355_.log) 2025-12-04T09:45:22.1803481Z 2025-12-04T09:45:22.1803869Z Finished inductor/test_cudacodecache 1/1 ... [2025-12-04 09:45:22.138578][2024.555050962], took 0.13min 2025-12-04T09:45:22.1804927Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cudacodecache/inductor.test_cudacodecache-c55849964bd60b32.xml 2025-12-04T09:45:22.6413790Z Uploading logs for 57118183167 to S3 2025-12-04T09:45:22.6875805Z Uploading artifacts took 0.46 seconds 2025-12-04T09:45:22.6876796Z inductor/test_cudacodecache 1/1 failed! 2025-12-04T09:45:22.6882180Z Running inductor/test_minifier_utils 1/1 ... [2025-12-04 09:45:22.687902][2025.104374083] 2025-12-04T09:45:22.6882810Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:45:22.6886973Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_minifier_utils.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:22.688319] 2025-12-04T09:45:26.2098745Z 2025-12-04T09:45:26.2099993Z inductor/test_minifier_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_minifier_utils_1.1_6733dac6651991a4_.log 2025-12-04T09:45:26.2100897Z Running 0 items in this shard: 2025-12-04T09:45:26.2101102Z 2025-12-04T09:45:26.2101476Z Finished inductor/test_minifier_utils 1/1 ... [2025-12-04 09:45:26.209513][2028.625986231], took 0.06min 2025-12-04T09:45:26.2162981Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_minifier_utils/inductor.test_minifier_utils-574f06f5728a4f68.xml 2025-12-04T09:45:26.2396588Z Running inductor/test_debug_trace 1/1 ... [2025-12-04 09:45:26.239263][2028.655736111] 2025-12-04T09:45:26.2397062Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:45:26.2399551Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_debug_trace.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:26.239570] 2025-12-04T09:45:33.0169665Z 2025-12-04T09:45:33.0170797Z inductor/test_debug_trace 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_debug_trace_1.1_97586d0a33b2ca76_.log 2025-12-04T09:45:33.0171639Z Running 0 items in this shard: 2025-12-04T09:45:33.0171832Z 2025-12-04T09:45:33.0172122Z Finished inductor/test_debug_trace 1/1 ... [2025-12-04 09:45:33.016551][2035.433024784], took 0.11min 2025-12-04T09:45:33.0232592Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_debug_trace/inductor.test_debug_trace-ffbc75868f825ab3.xml 2025-12-04T09:45:33.0902133Z Running export/test_tree_utils 1/1 ... [2025-12-04 09:45:33.089822][2035.506294909] 2025-12-04T09:45:33.0902707Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:45:33.0905559Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_tree_utils.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:33.090171] 2025-12-04T09:45:36.2610680Z 2025-12-04T09:45:36.2611790Z export/test_tree_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_tree_utils_1.1_f24f25a473c693f0_.log 2025-12-04T09:45:36.2612613Z Running 0 items in this shard: 2025-12-04T09:45:36.2612816Z 2025-12-04T09:45:36.2613086Z Finished export/test_tree_utils 1/1 ... [2025-12-04 09:45:36.260695][2038.677166209], took 0.05min 2025-12-04T09:45:36.2677471Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-7da9dd1b94d7565c.xml 2025-12-04T09:45:36.2942503Z Running inductor/test_triton_wrapper 1/1 ... [2025-12-04 09:45:36.293891][2038.710364404] 2025-12-04T09:45:36.2943376Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:45:36.2945872Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_triton_wrapper.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:36.294188] 2025-12-04T09:45:42.4708026Z 2025-12-04T09:45:42.4708850Z inductor/test_triton_wrapper 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_triton_wrapper_1.1_e94cfd406374d935_.log 2025-12-04T09:45:42.4709636Z Running 0 items in this shard: 2025-12-04T09:45:42.4709819Z 2025-12-04T09:45:42.4710133Z Finished inductor/test_triton_wrapper 1/1 ... [2025-12-04 09:45:42.470468][2044.88693938], took 0.10min 2025-12-04T09:45:42.4780283Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-0b0941e28d632943.xml 2025-12-04T09:45:42.5448865Z Running inductor/test_static_cuda_launcher 1/1 ... [2025-12-04 09:45:42.544507][2044.960980699] 2025-12-04T09:45:42.5449379Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:45:42.5451821Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_static_cuda_launcher.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:42.544826] 2025-12-04T09:45:48.7702138Z 2025-12-04T09:45:48.7703344Z inductor/test_static_cuda_launcher 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_static_cuda_launcher_1.1_d53c30cafc20ce47_.log 2025-12-04T09:45:48.7704941Z Running 0 items in this shard: 2025-12-04T09:45:48.7705134Z 2025-12-04T09:45:48.7705466Z Finished inductor/test_static_cuda_launcher 1/1 ... [2025-12-04 09:45:48.769884][2051.186357365], took 0.10min 2025-12-04T09:45:48.7775087Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-0c2b4708e9bf14de.xml 2025-12-04T09:45:48.8631648Z Running inductor/test_provenance_tracing 1/1 ... [2025-12-04 09:45:48.862742][2051.279214425] 2025-12-04T09:45:48.8632317Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:45:48.8634173Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_provenance_tracing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:48.863055] 2025-12-04T09:45:55.6402837Z 2025-12-04T09:45:55.6404062Z inductor/test_provenance_tracing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_provenance_tracing_1.1_870944fe6e309527_.log 2025-12-04T09:45:55.6404913Z Running 0 items in this shard: 2025-12-04T09:45:55.6405125Z 2025-12-04T09:45:55.6405430Z Finished inductor/test_provenance_tracing 1/1 ... [2025-12-04 09:45:55.639887][2058.0563602], took 0.11min 2025-12-04T09:45:55.6473842Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_provenance_tracing/inductor.test_provenance_tracing-e8a06af29d1ce55a.xml 2025-12-04T09:45:55.7249019Z Running inductor/test_memory_planning 1/1 ... [2025-12-04 09:45:55.724537][2058.141008684] 2025-12-04T09:45:55.7249683Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:45:55.7252541Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_memory_planning.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:55.724862] 2025-12-04T09:50:15.7581568Z 2025-12-04T09:50:15.7584705Z PRINTING LOG FILE of inductor/test_memory_planning 1/1 (test/test-reports/inductor.test_memory_planning_1.1_ab660c17d0398f84_.log) 2025-12-04T09:50:15.7586156Z Test results will be stored in test-reports/python-pytest/inductor.test_memory_planning/inductor.test_memory_planning-7b0b7b5cd4d35939.xml 2025-12-04T09:50:15.7587105Z ============================= test session starts ============================== 2025-12-04T09:50:15.7587907Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:50:15.7588726Z cachedir: .pytest_cache 2025-12-04T09:50:15.7589518Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:50:15.7590381Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:50:15.7590965Z configfile: pytest.ini 2025-12-04T09:50:15.7591800Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:50:15.7592736Z collecting ... collected 4 items 2025-12-04T09:50:15.7593193Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:50:15.7613762Z Running 50 items in this shard: test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint 2025-12-04T09:50:15.7634077Z 2025-12-04T09:50:15.7634537Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [7.0455s] [ 2%] 2025-12-04T09:50:15.7635776Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [5.2169s] [ 2%] 2025-12-04T09:50:15.7636808Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [5.2716s] [ 2%] 2025-12-04T09:50:15.7637822Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [5.2639s] [ 2%] 2025-12-04T09:50:15.7638833Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [5.3353s] [ 2%] 2025-12-04T09:50:15.7640024Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9617s] [ 2%] 2025-12-04T09:50:15.7641059Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9371s] [ 2%] 2025-12-04T09:50:15.7642064Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9493s] [ 2%] 2025-12-04T09:50:15.7643142Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9457s] [ 2%] 2025-12-04T09:50:15.7644161Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [5.1088s] [ 2%] 2025-12-04T09:50:15.7645183Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9426s] [ 2%] 2025-12-04T09:50:15.7646210Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9427s] [ 2%] 2025-12-04T09:50:15.7647209Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9504s] [ 2%] 2025-12-04T09:50:15.7648288Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9631s] [ 2%] 2025-12-04T09:50:15.7649325Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9668s] [ 2%] 2025-12-04T09:50:15.7650283Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9431s] [ 2%] 2025-12-04T09:50:15.7651334Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9541s] [ 2%] 2025-12-04T09:50:15.7652349Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9506s] [ 2%] 2025-12-04T09:50:15.7653435Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9997s] [ 2%] 2025-12-04T09:50:15.7654543Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9588s] [ 2%] 2025-12-04T09:50:15.7655594Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [5.0251s] [ 2%] 2025-12-04T09:50:15.7656639Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9615s] [ 2%] 2025-12-04T09:50:15.7657860Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9769s] [ 2%] 2025-12-04T09:50:15.7658911Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9645s] [ 2%] 2025-12-04T09:50:15.7659953Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9727s] [ 2%] 2025-12-04T09:50:15.7660977Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [5.0222s] [ 2%] 2025-12-04T09:50:15.7662016Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9671s] [ 2%] 2025-12-04T09:50:15.7663061Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9547s] [ 2%] 2025-12-04T09:50:15.7664070Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [5.1814s] [ 2%] 2025-12-04T09:50:15.7665198Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9482s] [ 2%] 2025-12-04T09:50:15.7685820Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9850s] [ 2%] 2025-12-04T09:50:15.7686933Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [5.0192s] [ 2%] 2025-12-04T09:50:15.7687778Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9513s] [ 2%] 2025-12-04T09:50:15.7688516Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9976s] [ 2%] 2025-12-04T09:50:15.7689256Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9584s] [ 2%] 2025-12-04T09:50:15.7689995Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9392s] [ 2%] 2025-12-04T09:50:15.7690741Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9643s] [ 2%] 2025-12-04T09:50:15.7691814Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9852s] [ 2%] 2025-12-04T09:50:15.7692876Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9623s] [ 2%] 2025-12-04T09:50:15.7693927Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9370s] [ 2%] 2025-12-04T09:50:15.7694972Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9368s] [ 2%] 2025-12-04T09:50:15.7696009Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9397s] [ 2%] 2025-12-04T09:50:15.7697053Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9458s] [ 2%] 2025-12-04T09:50:15.7697892Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9582s] [ 2%] 2025-12-04T09:50:15.7698635Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9567s] [ 2%] 2025-12-04T09:50:15.7699380Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint FAILED [4.9531s] [ 2%] 2025-12-04T09:50:15.7700133Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9297s] [ 2%] 2025-12-04T09:50:15.7700890Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9551s] [ 2%] 2025-12-04T09:50:15.7701631Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [5.2399s] [ 2%] 2025-12-04T09:50:15.7702372Z inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint PASSED [4.9719s] [ 2%] 2025-12-04T09:50:15.7702804Z 2025-12-04T09:50:15.7702933Z =================================== FAILURES =================================== 2025-12-04T09:50:15.7703383Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.7703820Z Traceback (most recent call last): 2025-12-04T09:50:15.7704399Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.7704963Z ).run(code) 2025-12-04T09:50:15.7705486Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.7705958Z Searched string: 2025-12-04T09:50:15.7706181Z int32_t u0_raw; 2025-12-04T09:50:15.7706510Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.7706878Z auto u0 = u0_raw; 2025-12-04T09:50:15.7707125Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.7707573Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7708188Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7708813Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7709206Z 2025-12-04T09:50:15.7709384Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7709747Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7710105Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.7710843Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.7711619Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.7711980Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7712326Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7712670Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.7713389Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.7714098Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.7714676Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.7715208Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.7715859Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.7716471Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.7716781Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.7717074Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.7717268Z 2025-12-04T09:50:15.7717272Z 2025-12-04T09:50:15.7717277Z 2025-12-04T09:50:15.7717280Z 2025-12-04T09:50:15.7717978Z Wrapper code written to: /tmp/tmpynmakc3o/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cvnxn3bizurhuzr7gkhjhxjjk4hyhhvhrzwo6jq6lrtygjt46rir.wrapper.cpp 2025-12-04T09:50:15.7719424Z Kernel code written to: /tmp/tmpynmakc3o/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cpixalai6zyzrpagcxp6pl7wnxudmwznsxpme4ci47jq24gmyg45.kernel.cpp 2025-12-04T09:50:15.7720355Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7720630Z 2025-12-04T09:50:15.7720634Z 2025-12-04T09:50:15.7720822Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.7721638Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.7722257Z 2025-12-04T09:50:15.7722487Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.7722993Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7723381Z unimplemented [] 2025-12-04T09:50:15.7723657Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7724468Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7725213Z graph_break [] 2025-12-04T09:50:15.7725524Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7726540Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7727469Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7727844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7728218Z unimplemented [] 2025-12-04T09:50:15.7728493Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7729378Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7730209Z graph_break [] 2025-12-04T09:50:15.7730527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7731487Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7732320Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7732725Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7733099Z unimplemented [] 2025-12-04T09:50:15.7733367Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7734252Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7735085Z graph_break [] 2025-12-04T09:50:15.7735747Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7736776Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7737605Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7737971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7738340Z unimplemented [] 2025-12-04T09:50:15.7738618Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7739513Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7740344Z graph_break [] 2025-12-04T09:50:15.7740647Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7741561Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7742412Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7742840Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.7743266Z Traceback (most recent call last): 2025-12-04T09:50:15.7743841Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.7744402Z ).run(code) 2025-12-04T09:50:15.7744819Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.7745305Z Searched string: 2025-12-04T09:50:15.7745525Z int32_t u0_raw; 2025-12-04T09:50:15.7745846Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.7746231Z auto u0 = u0_raw; 2025-12-04T09:50:15.7746480Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.7746924Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7747543Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7748330Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7748719Z 2025-12-04T09:50:15.7748858Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7749222Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7749551Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.7750295Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.7751077Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.7751424Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7751780Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7752125Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.7752854Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.7753560Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.7754087Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.7754618Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.7755254Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.7755856Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.7756168Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.7756456Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.7756644Z 2025-12-04T09:50:15.7756648Z 2025-12-04T09:50:15.7756652Z 2025-12-04T09:50:15.7756706Z 2025-12-04T09:50:15.7757385Z Wrapper code written to: /tmp/tmpw7pv0svu/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cgjwlds4j4fhy2xfxkgjbedsxcycy76i57x4bu7vps3eezkb7ayc.wrapper.cpp 2025-12-04T09:50:15.7758836Z Kernel code written to: /tmp/tmpw7pv0svu/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cbmy3vm5t27asmqclrs5sldgfynirjxyqx2gevp5suefbzlhnf23.kernel.cpp 2025-12-04T09:50:15.7759766Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7760030Z 2025-12-04T09:50:15.7760034Z 2025-12-04T09:50:15.7760223Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.7761029Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.7761650Z 2025-12-04T09:50:15.7761899Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.7762608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7763079Z unimplemented [] 2025-12-04T09:50:15.7763359Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7764231Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7764988Z graph_break [] 2025-12-04T09:50:15.7765308Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7766234Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7767072Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7767524Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7767904Z unimplemented [] 2025-12-04T09:50:15.7768172Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7769077Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7770051Z graph_break [] 2025-12-04T09:50:15.7770367Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7771291Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7772132Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7772507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7772882Z unimplemented [] 2025-12-04T09:50:15.7773208Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7774092Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7774979Z graph_break [] 2025-12-04T09:50:15.7775294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7776253Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7777138Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7777492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7777875Z unimplemented [] 2025-12-04T09:50:15.7778151Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7779044Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7779923Z graph_break [] 2025-12-04T09:50:15.7780242Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7781172Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7781993Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7782360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7782734Z unimplemented [] 2025-12-04T09:50:15.7783011Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7784163Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7785050Z graph_break [] 2025-12-04T09:50:15.7785362Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7786301Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7787137Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7787537Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.7787982Z Traceback (most recent call last): 2025-12-04T09:50:15.7788561Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.7789117Z ).run(code) 2025-12-04T09:50:15.7789550Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.7790026Z Searched string: 2025-12-04T09:50:15.7790246Z int32_t u0_raw; 2025-12-04T09:50:15.7790566Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.7790959Z auto u0 = u0_raw; 2025-12-04T09:50:15.7791226Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.7791679Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7792411Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7793073Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7793458Z 2025-12-04T09:50:15.7793597Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7793957Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7794289Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.7795034Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.7795812Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.7796164Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7796567Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7796911Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.7797577Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.7798292Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.7798828Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.7799352Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.7799985Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.7800598Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.7800911Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.7801254Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.7801445Z 2025-12-04T09:50:15.7801449Z 2025-12-04T09:50:15.7801453Z 2025-12-04T09:50:15.7801456Z 2025-12-04T09:50:15.7802137Z Wrapper code written to: /tmp/tmp3dda8dk0/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cupi7kwsy66cim4u25rel2co425mz3wnowd2vpxdktnqsbsu5zqf.wrapper.cpp 2025-12-04T09:50:15.7803592Z Kernel code written to: /tmp/tmp3dda8dk0/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/czs3f4ouyqoo7qfzf7xfiovltvjemtk4oyoudnzs7f5bbnk6nrsl.kernel.cpp 2025-12-04T09:50:15.7804550Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7804813Z 2025-12-04T09:50:15.7804817Z 2025-12-04T09:50:15.7805000Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.7805800Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.7806428Z 2025-12-04T09:50:15.7806651Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.7807169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7807633Z unimplemented [] 2025-12-04T09:50:15.7807906Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7808714Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7809463Z graph_break [] 2025-12-04T09:50:15.7809771Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7810684Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7811516Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7811875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7812259Z unimplemented [] 2025-12-04T09:50:15.7812541Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7814168Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7815015Z graph_break [] 2025-12-04T09:50:15.7815328Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7816251Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7817085Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7817455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7817830Z unimplemented [] 2025-12-04T09:50:15.7818107Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7819006Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7819883Z graph_break [] 2025-12-04T09:50:15.7820195Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7821118Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7821957Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7822314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7822694Z unimplemented [] 2025-12-04T09:50:15.7822973Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7823867Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7824975Z graph_break [] 2025-12-04T09:50:15.7825300Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7826324Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7827151Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7827517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7827898Z unimplemented [] 2025-12-04T09:50:15.7828176Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7829069Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7829907Z graph_break [] 2025-12-04T09:50:15.7830221Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7831138Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7831973Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7832341Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7832716Z unimplemented [] 2025-12-04T09:50:15.7832987Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7833884Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7834725Z graph_break [] 2025-12-04T09:50:15.7835035Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7836373Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7837213Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7837575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7837948Z unimplemented [] 2025-12-04T09:50:15.7838223Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7839123Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7839955Z graph_break [] 2025-12-04T09:50:15.7840259Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7841171Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7842080Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7842523Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.7842964Z Traceback (most recent call last): 2025-12-04T09:50:15.7843546Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.7844119Z ).run(code) 2025-12-04T09:50:15.7844544Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.7845036Z Searched string: 2025-12-04T09:50:15.7845268Z int32_t u0_raw; 2025-12-04T09:50:15.7845593Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.7845978Z auto u0 = u0_raw; 2025-12-04T09:50:15.7846236Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.7846761Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7847460Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7848102Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7848490Z 2025-12-04T09:50:15.7848632Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7849006Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7849349Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.7850097Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.7850885Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.7851250Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7851629Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7851991Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.7852673Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.7853390Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.7853940Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.7854481Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.7855128Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.7855750Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.7856081Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.7856393Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.7856593Z 2025-12-04T09:50:15.7856596Z 2025-12-04T09:50:15.7856600Z 2025-12-04T09:50:15.7856606Z 2025-12-04T09:50:15.7857401Z Wrapper code written to: /tmp/tmpxe3wnqs2/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/crpj7dz4k2hf7bipjm5e4dg4cbxzmks7bedsgusoswkzhijp2lci.wrapper.cpp 2025-12-04T09:50:15.7858862Z Kernel code written to: /tmp/tmpxe3wnqs2/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cc6edmxjttd2x3npllhthyultqraw3zsxaqevsnpgo3xoqne75wy.kernel.cpp 2025-12-04T09:50:15.7859813Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7860082Z 2025-12-04T09:50:15.7860086Z 2025-12-04T09:50:15.7860278Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.7861097Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.7861721Z 2025-12-04T09:50:15.7861948Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.7862475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7862912Z unimplemented [] 2025-12-04T09:50:15.7863220Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7864040Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7864804Z graph_break [] 2025-12-04T09:50:15.7865130Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7866047Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7866902Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7867278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7867664Z unimplemented [] 2025-12-04T09:50:15.7867987Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7868896Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7869737Z graph_break [] 2025-12-04T09:50:15.7870052Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7870981Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7871816Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7872179Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7872550Z unimplemented [] 2025-12-04T09:50:15.7872834Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7873773Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7874611Z graph_break [] 2025-12-04T09:50:15.7874929Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7875857Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7876690Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7877049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7877423Z unimplemented [] 2025-12-04T09:50:15.7877695Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7878575Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7879413Z graph_break [] 2025-12-04T09:50:15.7879731Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7880729Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7881549Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7881916Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7882296Z unimplemented [] 2025-12-04T09:50:15.7882573Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7883451Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7884289Z graph_break [] 2025-12-04T09:50:15.7884606Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7885561Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7886385Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7886755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7887141Z unimplemented [] 2025-12-04T09:50:15.7887484Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7888367Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7889193Z graph_break [] 2025-12-04T09:50:15.7889500Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7890483Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7891304Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7891666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7892033Z unimplemented [] 2025-12-04T09:50:15.7892321Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7893221Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7894051Z graph_break [] 2025-12-04T09:50:15.7894362Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7895292Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7896128Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7896493Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7896866Z unimplemented [] 2025-12-04T09:50:15.7897149Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7898050Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7898880Z graph_break [] 2025-12-04T09:50:15.7899201Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7900119Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7900955Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7901314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7901693Z unimplemented [] 2025-12-04T09:50:15.7901969Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7902988Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7903830Z graph_break [] 2025-12-04T09:50:15.7904139Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7916437Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7917277Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7917681Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.7918113Z Traceback (most recent call last): 2025-12-04T09:50:15.7918774Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 152, in test_unbacked_symint 2025-12-04T09:50:15.7919457Z ).check("RAIIAtenTensorHandle(tmp_tensor_handle_1);").run(code) 2025-12-04T09:50:15.7920365Z RuntimeError: Expected to find "AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_4, int_array_5, &tmp_tensor_handle_1));" but did not find it 2025-12-04T09:50:15.7921176Z Searched string: 2025-12-04T09:50:15.7921438Z AtenTensorHandle tmp_tensor_handle_0; 2025-12-04T09:50:15.7922099Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool1, 0, cached_torch_dtype_int32, 0, int_array_1, int_array_1, &tmp_tensor_handle_0)); 2025-12-04T09:50:15.7922847Z auto buf0 = RAIIAtenTensorHandle(tmp_tensor_handle_0); 2025-12-04T09:50:15.7923331Z // Topologically Sorted Source Nodes: [add], Original ATen: [aten.add] 2025-12-04T09:50:15.7923987Z call_triton_poi_fused_add_0(arg0_1, buf0, 1L, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.7924481Z arg0_1.reset(); 2025-12-04T09:50:15.7924700Z int32_t u0_raw; 2025-12-04T09:50:15.7925044Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.7925434Z auto u0 = u0_raw; 2025-12-04T09:50:15.7925678Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.7926137Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7926868Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7927618Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.7928001Z 2025-12-04T09:50:15.7928134Z const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.7928501Z const int64_t int_array_3[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.7928846Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.7929600Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_2, int_array_3, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.7930387Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.7930740Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.7931402Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_2, int_array_3, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.7932108Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.7932700Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.7933237Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.7933880Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.7934490Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.7934815Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.7935107Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.7935794Z 2025-12-04T09:50:15.7935799Z 2025-12-04T09:50:15.7935803Z 2025-12-04T09:50:15.7935807Z 2025-12-04T09:50:15.7936496Z Wrapper code written to: /tmp/tmp8jct58wp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/chl47nn5sum3xy4ll77za6csl3dyhq2qf7k4l34y22amiydicuos.wrapper.cpp 2025-12-04T09:50:15.7937914Z Kernel code written to: /tmp/tmp8jct58wp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/czvwj7hef5hiutsegg3svscv3l3zfgefsx5senos3t2jm7jj2idu.kernel.cpp 2025-12-04T09:50:15.7939160Z From CHECK: AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_4, int_array_5, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.7939745Z 2025-12-04T09:50:15.7939749Z 2025-12-04T09:50:15.7939935Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.7940810Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.7941442Z 2025-12-04T09:50:15.7941666Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.7942182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7942564Z unimplemented [] 2025-12-04T09:50:15.7942840Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7943645Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7944402Z graph_break [] 2025-12-04T09:50:15.7944713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7945640Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7946532Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7946902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7947277Z unimplemented [] 2025-12-04T09:50:15.7947545Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7948435Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7949268Z graph_break [] 2025-12-04T09:50:15.7949580Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7950494Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7951335Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7951702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7952072Z unimplemented [] 2025-12-04T09:50:15.7952356Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7953300Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7954141Z graph_break [] 2025-12-04T09:50:15.7954449Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7955369Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7956202Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7956563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7956942Z unimplemented [] 2025-12-04T09:50:15.7957219Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7958198Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7959030Z graph_break [] 2025-12-04T09:50:15.7959339Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7960256Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7961090Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7961442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7961824Z unimplemented [] 2025-12-04T09:50:15.7962106Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7963040Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7963923Z graph_break [] 2025-12-04T09:50:15.7964231Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7965145Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7965962Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7966326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7966705Z unimplemented [] 2025-12-04T09:50:15.7966984Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7967984Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7969544Z graph_break [] 2025-12-04T09:50:15.7969858Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7970763Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7971583Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7971940Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7972314Z unimplemented [] 2025-12-04T09:50:15.7972582Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7973519Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7974352Z graph_break [] 2025-12-04T09:50:15.7974655Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7975565Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7976392Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7976751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7977122Z unimplemented [] 2025-12-04T09:50:15.7977394Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7978281Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7979112Z graph_break [] 2025-12-04T09:50:15.7979416Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7980428Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7981259Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7981618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7981986Z unimplemented [] 2025-12-04T09:50:15.7982259Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7983154Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7983981Z graph_break [] 2025-12-04T09:50:15.7984287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7985199Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7986079Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7986440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7986812Z unimplemented [] 2025-12-04T09:50:15.7987087Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7987975Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7988817Z graph_break [] 2025-12-04T09:50:15.7989130Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7990049Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7990935Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7991293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7991677Z unimplemented [] 2025-12-04T09:50:15.7991942Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7992833Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7993715Z graph_break [] 2025-12-04T09:50:15.7994022Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7994933Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.7995766Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.7996137Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.7996516Z unimplemented [] 2025-12-04T09:50:15.7996782Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.7997695Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.7998530Z graph_break [] 2025-12-04T09:50:15.7998835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.7999754Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8000577Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8000974Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8001399Z Traceback (most recent call last): 2025-12-04T09:50:15.8001976Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8002534Z ).run(code) 2025-12-04T09:50:15.8003060Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8003559Z Searched string: 2025-12-04T09:50:15.8003784Z int32_t u0_raw; 2025-12-04T09:50:15.8004124Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8004498Z auto u0 = u0_raw; 2025-12-04T09:50:15.8004749Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8005192Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8005796Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8006439Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8006885Z 2025-12-04T09:50:15.8007016Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8007435Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8007769Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8008515Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8009292Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8009653Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8010004Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8010352Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8011027Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8011785Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8012332Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8012875Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8013531Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8014128Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8014445Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8014742Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8014933Z 2025-12-04T09:50:15.8014937Z 2025-12-04T09:50:15.8014941Z 2025-12-04T09:50:15.8014944Z 2025-12-04T09:50:15.8015628Z Wrapper code written to: /tmp/tmpus4s7msp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c32rinomxhjvwr4rzfv5wrndcnn2avny2ug5lapit3q3lecrwi2u.wrapper.cpp 2025-12-04T09:50:15.8017069Z Kernel code written to: /tmp/tmpus4s7msp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cxisfpmlgnp3b7qua4c5xvnmkaejdeklhn23wbpjytiwkgdzqp4i.kernel.cpp 2025-12-04T09:50:15.8017993Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8018256Z 2025-12-04T09:50:15.8018264Z 2025-12-04T09:50:15.8018443Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8019254Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8019868Z 2025-12-04T09:50:15.8020093Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8020592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8020967Z unimplemented [] 2025-12-04T09:50:15.8021246Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8022050Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8022844Z graph_break [] 2025-12-04T09:50:15.8023248Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8024170Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8025005Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8025362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8025737Z unimplemented [] 2025-12-04T09:50:15.8026016Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8026891Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8027721Z graph_break [] 2025-12-04T09:50:15.8028028Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8028993Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8029817Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8030186Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8030558Z unimplemented [] 2025-12-04T09:50:15.8030827Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8031709Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8032536Z graph_break [] 2025-12-04T09:50:15.8032844Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8033797Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8034753Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8035209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8035783Z unimplemented [] 2025-12-04T09:50:15.8036050Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8036942Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8037781Z graph_break [] 2025-12-04T09:50:15.8038086Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8039002Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8039836Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8040207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8040575Z unimplemented [] 2025-12-04T09:50:15.8040850Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8041735Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8042558Z graph_break [] 2025-12-04T09:50:15.8042870Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8043780Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8044607Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8044964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8045342Z unimplemented [] 2025-12-04T09:50:15.8045778Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8046674Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8047603Z graph_break [] 2025-12-04T09:50:15.8047916Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8048829Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8049648Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8050017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8050399Z unimplemented [] 2025-12-04T09:50:15.8050752Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8051644Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8052529Z graph_break [] 2025-12-04T09:50:15.8052840Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8053755Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8054584Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8054944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8055327Z unimplemented [] 2025-12-04T09:50:15.8055603Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8056574Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8057413Z graph_break [] 2025-12-04T09:50:15.8057732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8058646Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8059479Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8059842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8060211Z unimplemented [] 2025-12-04T09:50:15.8060482Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8061369Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8062219Z graph_break [] 2025-12-04T09:50:15.8062524Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8063462Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8064303Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8064662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8065037Z unimplemented [] 2025-12-04T09:50:15.8065312Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8066204Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8067035Z graph_break [] 2025-12-04T09:50:15.8067348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8068345Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8069179Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8069534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8069912Z unimplemented [] 2025-12-04T09:50:15.8070188Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8071066Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8071898Z graph_break [] 2025-12-04T09:50:15.8072212Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8073210Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8074026Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8074394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8074772Z unimplemented [] 2025-12-04T09:50:15.8075041Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8075940Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8076781Z graph_break [] 2025-12-04T09:50:15.8077091Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8078004Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8078895Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8079264Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8079647Z unimplemented [] 2025-12-04T09:50:15.8079924Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8080827Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8081668Z graph_break [] 2025-12-04T09:50:15.8081982Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8082961Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8083800Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8084181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8084562Z unimplemented [] 2025-12-04T09:50:15.8084862Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8085771Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8086613Z graph_break [] 2025-12-04T09:50:15.8086929Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8087923Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8088777Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8089188Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8089647Z Traceback (most recent call last): 2025-12-04T09:50:15.8090318Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8090901Z ).run(code) 2025-12-04T09:50:15.8091331Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8091828Z Searched string: 2025-12-04T09:50:15.8092060Z int32_t u0_raw; 2025-12-04T09:50:15.8092414Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8092835Z auto u0 = u0_raw; 2025-12-04T09:50:15.8093092Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8093549Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8094173Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8094833Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8095265Z 2025-12-04T09:50:15.8095408Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8095784Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8096122Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8096874Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8097668Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8098028Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8098396Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8098757Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8099434Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8100192Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8100762Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8101309Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8101973Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8102601Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8102938Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8103240Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8103431Z 2025-12-04T09:50:15.8103434Z 2025-12-04T09:50:15.8103438Z 2025-12-04T09:50:15.8103442Z 2025-12-04T09:50:15.8104119Z Wrapper code written to: /tmp/tmpo1v55153/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cuvl6siq5dkkznzwmgn2eken45qtj6syxgc5feulqbhinj5yo5y3.wrapper.cpp 2025-12-04T09:50:15.8105571Z Kernel code written to: /tmp/tmpo1v55153/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cgalohs3wmwvvu37brmwwzxo2ep75c73kqjgmlnxadvlio3itq24.kernel.cpp 2025-12-04T09:50:15.8106519Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8106787Z 2025-12-04T09:50:15.8106791Z 2025-12-04T09:50:15.8106990Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8107805Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8108450Z 2025-12-04T09:50:15.8108685Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8109210Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8109603Z unimplemented [] 2025-12-04T09:50:15.8109880Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8110788Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8111563Z graph_break [] 2025-12-04T09:50:15.8111983Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8112929Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8113824Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8114198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8114584Z unimplemented [] 2025-12-04T09:50:15.8114883Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8115787Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8116670Z graph_break [] 2025-12-04T09:50:15.8116976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8117926Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8118766Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8119147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8119522Z unimplemented [] 2025-12-04T09:50:15.8119808Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8120721Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8121569Z graph_break [] 2025-12-04T09:50:15.8121933Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8122934Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8123774Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8124134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8124519Z unimplemented [] 2025-12-04T09:50:15.8124807Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8125711Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8126560Z graph_break [] 2025-12-04T09:50:15.8126895Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8127894Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8128728Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8129109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8129497Z unimplemented [] 2025-12-04T09:50:15.8129789Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8130685Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8131554Z graph_break [] 2025-12-04T09:50:15.8131871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8132859Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8133711Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8134092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8134582Z unimplemented [] 2025-12-04T09:50:15.8134868Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8136285Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8137129Z graph_break [] 2025-12-04T09:50:15.8137444Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8138380Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8139226Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8139606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8151992Z unimplemented [] 2025-12-04T09:50:15.8152316Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8153258Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8154080Z graph_break [] 2025-12-04T09:50:15.8154384Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8155299Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8156113Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8156461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8156941Z unimplemented [] 2025-12-04T09:50:15.8157203Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8158086Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8158913Z graph_break [] 2025-12-04T09:50:15.8159225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8160142Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8160971Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8161339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8161729Z unimplemented [] 2025-12-04T09:50:15.8162003Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8162897Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8163737Z graph_break [] 2025-12-04T09:50:15.8164058Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8164973Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8165804Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8166173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8166547Z unimplemented [] 2025-12-04T09:50:15.8166820Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8167763Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8168602Z graph_break [] 2025-12-04T09:50:15.8168901Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8170001Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8170825Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8171184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8171551Z unimplemented [] 2025-12-04T09:50:15.8171827Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8172718Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8173547Z graph_break [] 2025-12-04T09:50:15.8173851Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8174819Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8175645Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8176000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8176378Z unimplemented [] 2025-12-04T09:50:15.8176657Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8177539Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8178369Z graph_break [] 2025-12-04T09:50:15.8178680Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8179649Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8180472Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8180828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8181204Z unimplemented [] 2025-12-04T09:50:15.8181486Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8182391Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8183255Z graph_break [] 2025-12-04T09:50:15.8183573Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8184486Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8185314Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8185677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8186048Z unimplemented [] 2025-12-04T09:50:15.8186314Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8187196Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8188036Z graph_break [] 2025-12-04T09:50:15.8188348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8189258Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8190087Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8190447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8190813Z unimplemented [] 2025-12-04T09:50:15.8191169Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8192061Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8192887Z graph_break [] 2025-12-04T09:50:15.8193198Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8194111Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8194950Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8195354Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8195776Z Traceback (most recent call last): 2025-12-04T09:50:15.8196398Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8196973Z ).run(code) 2025-12-04T09:50:15.8197392Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8197870Z Searched string: 2025-12-04T09:50:15.8198094Z int32_t u0_raw; 2025-12-04T09:50:15.8198414Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8198794Z auto u0 = u0_raw; 2025-12-04T09:50:15.8199044Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8199495Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8200099Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8200744Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8201174Z 2025-12-04T09:50:15.8201317Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8201686Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8202017Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8202758Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8203535Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8203886Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8204235Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8204578Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8205249Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8205957Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8206496Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8207038Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8207780Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8208397Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8208724Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8209015Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8209199Z 2025-12-04T09:50:15.8209203Z 2025-12-04T09:50:15.8209207Z 2025-12-04T09:50:15.8209210Z 2025-12-04T09:50:15.8209889Z Wrapper code written to: /tmp/tmpojoo8738/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/czglehddruugo4yh3rb2m74ahflahffybpnjoz4vnhomsq4fdi3l.wrapper.cpp 2025-12-04T09:50:15.8211456Z Kernel code written to: /tmp/tmpojoo8738/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cygljwezmp44v7se3tjmyf3sk3ugun22b2uuh2smv557py6opwk6.kernel.cpp 2025-12-04T09:50:15.8212397Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8212788Z 2025-12-04T09:50:15.8212793Z 2025-12-04T09:50:15.8212992Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8213805Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8214426Z 2025-12-04T09:50:15.8214655Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8215178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8215559Z unimplemented [] 2025-12-04T09:50:15.8215839Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8216655Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8217462Z graph_break [] 2025-12-04T09:50:15.8217790Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8218720Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8219566Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8219947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8220328Z unimplemented [] 2025-12-04T09:50:15.8220605Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8221512Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8222419Z graph_break [] 2025-12-04T09:50:15.8222770Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8223701Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8224537Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8224908Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8225280Z unimplemented [] 2025-12-04T09:50:15.8225566Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8226478Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8227312Z graph_break [] 2025-12-04T09:50:15.8227639Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8228583Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8229435Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8229788Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8230165Z unimplemented [] 2025-12-04T09:50:15.8230443Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8231342Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8232166Z graph_break [] 2025-12-04T09:50:15.8232480Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8233446Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8234270Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8234710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8235090Z unimplemented [] 2025-12-04T09:50:15.8235711Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8236593Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8237430Z graph_break [] 2025-12-04T09:50:15.8237740Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8238649Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8239478Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8239933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8240318Z unimplemented [] 2025-12-04T09:50:15.8240598Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8241484Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8242318Z graph_break [] 2025-12-04T09:50:15.8242636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8243596Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8244424Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8244786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8245229Z unimplemented [] 2025-12-04T09:50:15.8245513Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8246416Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8247304Z graph_break [] 2025-12-04T09:50:15.8247607Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8248521Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8249350Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8249716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8250081Z unimplemented [] 2025-12-04T09:50:15.8250362Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8251261Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8252082Z graph_break [] 2025-12-04T09:50:15.8252394Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8253309Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8254133Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8254496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8254868Z unimplemented [] 2025-12-04T09:50:15.8255148Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8256021Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8256857Z graph_break [] 2025-12-04T09:50:15.8257316Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8258233Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8259042Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8259407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8259778Z unimplemented [] 2025-12-04T09:50:15.8260056Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8260945Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8261772Z graph_break [] 2025-12-04T09:50:15.8262122Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8263100Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8263924Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8264278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8264656Z unimplemented [] 2025-12-04T09:50:15.8264931Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8265816Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8266643Z graph_break [] 2025-12-04T09:50:15.8266999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8267931Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8268751Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8269111Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8269491Z unimplemented [] 2025-12-04T09:50:15.8269764Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8270639Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8271467Z graph_break [] 2025-12-04T09:50:15.8271780Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8272740Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8273566Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8273935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8274317Z unimplemented [] 2025-12-04T09:50:15.8274586Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8275492Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8276334Z graph_break [] 2025-12-04T09:50:15.8276654Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8277555Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8278388Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8278765Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8279137Z unimplemented [] 2025-12-04T09:50:15.8279500Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8280390Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8281216Z graph_break [] 2025-12-04T09:50:15.8281517Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8282447Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8283318Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8283684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8284088Z unimplemented [] 2025-12-04T09:50:15.8284375Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8285275Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8286105Z graph_break [] 2025-12-04T09:50:15.8286412Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8287382Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8288219Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8288578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8289006Z unimplemented [] 2025-12-04T09:50:15.8289290Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8290183Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8291006Z graph_break [] 2025-12-04T09:50:15.8291305Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8292236Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8293097Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8293503Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8293940Z Traceback (most recent call last): 2025-12-04T09:50:15.8294528Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8295083Z ).run(code) 2025-12-04T09:50:15.8295521Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8296008Z Searched string: 2025-12-04T09:50:15.8296230Z int32_t u0_raw; 2025-12-04T09:50:15.8296562Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8296937Z auto u0 = u0_raw; 2025-12-04T09:50:15.8297172Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8297621Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8298235Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8298864Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8299244Z 2025-12-04T09:50:15.8299373Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8299732Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8300064Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8300891Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8301657Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8302014Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8302364Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8302700Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8303409Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8303571Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8303842Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8304042Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8304439Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8304545Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8304649Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8304756Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8304761Z 2025-12-04T09:50:15.8304766Z 2025-12-04T09:50:15.8304769Z 2025-12-04T09:50:15.8304773Z 2025-12-04T09:50:15.8305454Z Wrapper code written to: /tmp/tmpldcbptk5/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cb6bbvswsmkgcpryj77u7wue5dmvexmfbg52j4qaepk5ljzqq3jf.wrapper.cpp 2025-12-04T09:50:15.8306116Z Kernel code written to: /tmp/tmpldcbptk5/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c2jafjcq5hjdfypuqff7zhhjgnfo7h247dkgrv44qnjxi2bqxbq2.kernel.cpp 2025-12-04T09:50:15.8306315Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8306322Z 2025-12-04T09:50:15.8306326Z 2025-12-04T09:50:15.8306511Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8307035Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8307040Z 2025-12-04T09:50:15.8307259Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8307436Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8307521Z unimplemented [] 2025-12-04T09:50:15.8307660Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8308229Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8308317Z graph_break [] 2025-12-04T09:50:15.8308502Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8309139Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8309242Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8309413Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8309499Z unimplemented [] 2025-12-04T09:50:15.8309634Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8310274Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8310358Z graph_break [] 2025-12-04T09:50:15.8310539Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8311165Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8311349Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8311523Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8311607Z unimplemented [] 2025-12-04T09:50:15.8311744Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8312394Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8312476Z graph_break [] 2025-12-04T09:50:15.8312654Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8313325Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8313466Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8313637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8313731Z unimplemented [] 2025-12-04T09:50:15.8313868Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8314510Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8314596Z graph_break [] 2025-12-04T09:50:15.8314766Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8315390Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8315495Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8315742Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8315832Z unimplemented [] 2025-12-04T09:50:15.8315972Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8316621Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8316710Z graph_break [] 2025-12-04T09:50:15.8316884Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8317515Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8317620Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8317794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8317883Z unimplemented [] 2025-12-04T09:50:15.8318031Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8318683Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8318772Z graph_break [] 2025-12-04T09:50:15.8318941Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8319566Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8319673Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8319845Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8319935Z unimplemented [] 2025-12-04T09:50:15.8320067Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8320716Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8320888Z graph_break [] 2025-12-04T09:50:15.8321064Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8321689Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8321799Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8321970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8322061Z unimplemented [] 2025-12-04T09:50:15.8322191Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8322833Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8322964Z graph_break [] 2025-12-04T09:50:15.8323142Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8323777Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8323875Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8324051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8324143Z unimplemented [] 2025-12-04T09:50:15.8324272Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8324912Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8325045Z graph_break [] 2025-12-04T09:50:15.8325217Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8325857Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8325953Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8326122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8326212Z unimplemented [] 2025-12-04T09:50:15.8326341Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8326982Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8327077Z graph_break [] 2025-12-04T09:50:15.8327301Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8327939Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8328045Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8328223Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8328320Z unimplemented [] 2025-12-04T09:50:15.8328448Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8329104Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8329185Z graph_break [] 2025-12-04T09:50:15.8329354Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8329984Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8330088Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8330339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8330434Z unimplemented [] 2025-12-04T09:50:15.8330564Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8331208Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8331290Z graph_break [] 2025-12-04T09:50:15.8331463Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8332090Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8332190Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8332408Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8332493Z unimplemented [] 2025-12-04T09:50:15.8332627Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8333282Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8333364Z graph_break [] 2025-12-04T09:50:15.8333536Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8334165Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8334264Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8334442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8334579Z unimplemented [] 2025-12-04T09:50:15.8334710Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8335587Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8335675Z graph_break [] 2025-12-04T09:50:15.8335848Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8336483Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8336579Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8336756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8336841Z unimplemented [] 2025-12-04T09:50:15.8336979Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8337628Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8337714Z graph_break [] 2025-12-04T09:50:15.8337892Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8338515Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8338610Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8338784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8338875Z unimplemented [] 2025-12-04T09:50:15.8339004Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8339656Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8339740Z graph_break [] 2025-12-04T09:50:15.8340055Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8340682Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8340778Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8340954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8341037Z unimplemented [] 2025-12-04T09:50:15.8341168Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8341813Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8341897Z graph_break [] 2025-12-04T09:50:15.8342132Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8342817Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8342914Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8343139Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8343246Z Traceback (most recent call last): 2025-12-04T09:50:15.8343633Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8343713Z ).run(code) 2025-12-04T09:50:15.8344005Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8344095Z Searched string: 2025-12-04T09:50:15.8344238Z int32_t u0_raw; 2025-12-04T09:50:15.8344428Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8344516Z auto u0 = u0_raw; 2025-12-04T09:50:15.8344612Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8344885Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8345136Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8345415Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8345420Z 2025-12-04T09:50:15.8345554Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8345683Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8345790Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8346331Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8346464Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8346601Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8346728Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8346845Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8347297Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8347453Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8347730Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8347877Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8348260Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8348377Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8348485Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8348586Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8348590Z 2025-12-04T09:50:15.8348678Z 2025-12-04T09:50:15.8348682Z 2025-12-04T09:50:15.8348686Z 2025-12-04T09:50:15.8349357Z Wrapper code written to: /tmp/tmpzy568zuz/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c6ahe4lfy2mnl5l36m23pjwmp7cwevf3c4hjgspxzeyevquk733e.wrapper.cpp 2025-12-04T09:50:15.8350017Z Kernel code written to: /tmp/tmpzy568zuz/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/caoclm3q33lmmn5c6i4rhj6gnh64s3lxnggraykdmhbttsyvdgff.kernel.cpp 2025-12-04T09:50:15.8350184Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8350189Z 2025-12-04T09:50:15.8350193Z 2025-12-04T09:50:15.8350371Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8350888Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8350959Z 2025-12-04T09:50:15.8351180Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8351363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8351455Z unimplemented [] 2025-12-04T09:50:15.8351589Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8352157Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8352239Z graph_break [] 2025-12-04T09:50:15.8352426Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8353105Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8353243Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8353431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8353518Z unimplemented [] 2025-12-04T09:50:15.8353659Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8354311Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8354391Z graph_break [] 2025-12-04T09:50:15.8354566Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8355195Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8355290Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8355469Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8355559Z unimplemented [] 2025-12-04T09:50:15.8355691Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8356347Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8356425Z graph_break [] 2025-12-04T09:50:15.8356596Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8357223Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8357316Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8357497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8357584Z unimplemented [] 2025-12-04T09:50:15.8357717Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8358444Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8358527Z graph_break [] 2025-12-04T09:50:15.8358708Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8359333Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8359431Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8359611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8359694Z unimplemented [] 2025-12-04T09:50:15.8359823Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8360477Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8360598Z graph_break [] 2025-12-04T09:50:15.8360779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8361403Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8361500Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8361683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8361767Z unimplemented [] 2025-12-04T09:50:15.8361896Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8362547Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8362671Z graph_break [] 2025-12-04T09:50:15.8362848Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8363524Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8363618Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8363798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8368977Z unimplemented [] 2025-12-04T09:50:15.8369138Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8369789Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8369884Z graph_break [] 2025-12-04T09:50:15.8370072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8370716Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8370815Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8370992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8371082Z unimplemented [] 2025-12-04T09:50:15.8371216Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8371859Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8371944Z graph_break [] 2025-12-04T09:50:15.8372117Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8372748Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8372849Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8373164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8373274Z unimplemented [] 2025-12-04T09:50:15.8373405Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8374057Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8374139Z graph_break [] 2025-12-04T09:50:15.8374311Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8374933Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8375071Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8375244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8375342Z unimplemented [] 2025-12-04T09:50:15.8375472Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8376116Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8376197Z graph_break [] 2025-12-04T09:50:15.8376369Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8377000Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8377096Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8377323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8377412Z unimplemented [] 2025-12-04T09:50:15.8377554Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8378205Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8378290Z graph_break [] 2025-12-04T09:50:15.8378468Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8379097Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8379193Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8379374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8379472Z unimplemented [] 2025-12-04T09:50:15.8379604Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8380268Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8380355Z graph_break [] 2025-12-04T09:50:15.8380530Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8381156Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8381255Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8381442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8381527Z unimplemented [] 2025-12-04T09:50:15.8381657Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8382315Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8382481Z graph_break [] 2025-12-04T09:50:15.8382664Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8383288Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8383384Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8383569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8383655Z unimplemented [] 2025-12-04T09:50:15.8383787Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8384428Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8384553Z graph_break [] 2025-12-04T09:50:15.8384734Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8385363Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8385462Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8385645Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8385734Z unimplemented [] 2025-12-04T09:50:15.8385870Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8386527Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8386656Z graph_break [] 2025-12-04T09:50:15.8386830Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8387462Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8387559Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8387739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8387825Z unimplemented [] 2025-12-04T09:50:15.8387970Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8388610Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8388694Z graph_break [] 2025-12-04T09:50:15.8388880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8389503Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8389613Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8389783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8389869Z unimplemented [] 2025-12-04T09:50:15.8390008Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8390645Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8390725Z graph_break [] 2025-12-04T09:50:15.8390904Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8391524Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8391628Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8391903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8391990Z unimplemented [] 2025-12-04T09:50:15.8392127Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8392770Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8392849Z graph_break [] 2025-12-04T09:50:15.8393021Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8393688Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8393793Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8394004Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8394146Z Traceback (most recent call last): 2025-12-04T09:50:15.8394538Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8394615Z ).run(code) 2025-12-04T09:50:15.8394904Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8394991Z Searched string: 2025-12-04T09:50:15.8395070Z int32_t u0_raw; 2025-12-04T09:50:15.8395267Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8395349Z auto u0 = u0_raw; 2025-12-04T09:50:15.8395447Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8395717Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8395969Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8396296Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8396302Z 2025-12-04T09:50:15.8396435Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8396559Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8396669Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8397205Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8397337Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8397474Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8397597Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8397713Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8398151Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8398315Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8398596Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8398743Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8399135Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8399240Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8399343Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8399450Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8399454Z 2025-12-04T09:50:15.8399458Z 2025-12-04T09:50:15.8399462Z 2025-12-04T09:50:15.8399466Z 2025-12-04T09:50:15.8400143Z Wrapper code written to: /tmp/tmpw2qoh7sg/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cjpjdca5j24wpvshtns3kzrfb2ssk6aorqlrbhz32uxsyism3epr.wrapper.cpp 2025-12-04T09:50:15.8400886Z Kernel code written to: /tmp/tmpw2qoh7sg/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/csmsledbnqmvnyxfs3la7tloykqs6vbf5ijravu3atekvmx3vgsr.kernel.cpp 2025-12-04T09:50:15.8401049Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8401053Z 2025-12-04T09:50:15.8401057Z 2025-12-04T09:50:15.8401236Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8401758Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8401762Z 2025-12-04T09:50:15.8401983Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8402173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8402261Z unimplemented [] 2025-12-04T09:50:15.8402395Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8403007Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8403097Z graph_break [] 2025-12-04T09:50:15.8403302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8403951Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8404048Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8404223Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8404307Z unimplemented [] 2025-12-04T09:50:15.8404446Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8405087Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8405210Z graph_break [] 2025-12-04T09:50:15.8405390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8406016Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8406115Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8406292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8406379Z unimplemented [] 2025-12-04T09:50:15.8406509Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8407155Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8407291Z graph_break [] 2025-12-04T09:50:15.8407471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8408098Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8408192Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8408370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8408456Z unimplemented [] 2025-12-04T09:50:15.8408590Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8409231Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8409313Z graph_break [] 2025-12-04T09:50:15.8409494Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8410196Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8410298Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8410473Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8410555Z unimplemented [] 2025-12-04T09:50:15.8410694Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8411339Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8411426Z graph_break [] 2025-12-04T09:50:15.8411604Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8412219Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8412364Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8412540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8412628Z unimplemented [] 2025-12-04T09:50:15.8412767Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8413413Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8413498Z graph_break [] 2025-12-04T09:50:15.8413677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8414298Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8414447Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8414623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8414708Z unimplemented [] 2025-12-04T09:50:15.8414845Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8415484Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8415574Z graph_break [] 2025-12-04T09:50:15.8415747Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8416370Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8416467Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8416635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8416725Z unimplemented [] 2025-12-04T09:50:15.8416858Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8417507Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8417595Z graph_break [] 2025-12-04T09:50:15.8417765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8418388Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8418488Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8418660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8418746Z unimplemented [] 2025-12-04T09:50:15.8418882Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8419604Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8419692Z graph_break [] 2025-12-04T09:50:15.8419862Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8420481Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8420579Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8420750Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8420841Z unimplemented [] 2025-12-04T09:50:15.8420970Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8421613Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8421741Z graph_break [] 2025-12-04T09:50:15.8421923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8422545Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8422642Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8422811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8422898Z unimplemented [] 2025-12-04T09:50:15.8423027Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8423670Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8423798Z graph_break [] 2025-12-04T09:50:15.8423973Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8424604Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8424698Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8424869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8424958Z unimplemented [] 2025-12-04T09:50:15.8425087Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8425729Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8425814Z graph_break [] 2025-12-04T09:50:15.8425984Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8426618Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8426713Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8426886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8426984Z unimplemented [] 2025-12-04T09:50:15.8427114Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8427755Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8427845Z graph_break [] 2025-12-04T09:50:15.8428022Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8428649Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8428749Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8429021Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8429113Z unimplemented [] 2025-12-04T09:50:15.8429246Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8429886Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8429967Z graph_break [] 2025-12-04T09:50:15.8430139Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8430766Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8430904Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8431074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8431173Z unimplemented [] 2025-12-04T09:50:15.8431300Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8431952Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8432034Z graph_break [] 2025-12-04T09:50:15.8432202Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8432827Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8432926Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8433141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8433228Z unimplemented [] 2025-12-04T09:50:15.8433357Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8434008Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8434093Z graph_break [] 2025-12-04T09:50:15.8434267Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8434892Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8434986Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8435163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8435488Z unimplemented [] 2025-12-04T09:50:15.8435676Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8436351Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8436440Z graph_break [] 2025-12-04T09:50:15.8436609Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8437232Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8437325Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8437497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8437584Z unimplemented [] 2025-12-04T09:50:15.8437712Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8438357Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8438443Z graph_break [] 2025-12-04T09:50:15.8438759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8439379Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8439478Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8439659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8439744Z unimplemented [] 2025-12-04T09:50:15.8439873Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8440517Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8440657Z graph_break [] 2025-12-04T09:50:15.8440833Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8441459Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8441552Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8441766Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8441869Z Traceback (most recent call last): 2025-12-04T09:50:15.8442241Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8442324Z ).run(code) 2025-12-04T09:50:15.8442617Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8442764Z Searched string: 2025-12-04T09:50:15.8442842Z int32_t u0_raw; 2025-12-04T09:50:15.8443031Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8443118Z auto u0 = u0_raw; 2025-12-04T09:50:15.8443220Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8443475Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8443725Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8443998Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8444003Z 2025-12-04T09:50:15.8444133Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8444260Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8444363Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8444902Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8445033Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8445164Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8445284Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8445397Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8445833Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8445989Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8446259Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8446409Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8446792Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8446910Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8447011Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8447191Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8447196Z 2025-12-04T09:50:15.8447200Z 2025-12-04T09:50:15.8447204Z 2025-12-04T09:50:15.8447211Z 2025-12-04T09:50:15.8447938Z Wrapper code written to: /tmp/tmpvxygh32y/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cgpl4og5ixv4rtug3nalbf5qwfe36jepzefotlyqiwruuiefqv2u.wrapper.cpp 2025-12-04T09:50:15.8448600Z Kernel code written to: /tmp/tmpvxygh32y/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cwopjd5pu2bkgij4beg677lnpvemwrpe5a7bmotya6popxocixzo.kernel.cpp 2025-12-04T09:50:15.8448771Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8448775Z 2025-12-04T09:50:15.8448779Z 2025-12-04T09:50:15.8448964Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8449489Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8449534Z 2025-12-04T09:50:15.8449761Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8449936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8450023Z unimplemented [] 2025-12-04T09:50:15.8450157Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8450727Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8450813Z graph_break [] 2025-12-04T09:50:15.8450984Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8451616Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8451753Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8451932Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8452022Z unimplemented [] 2025-12-04T09:50:15.8452152Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8452798Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8452884Z graph_break [] 2025-12-04T09:50:15.8453054Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8453677Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8453776Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8453959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8454050Z unimplemented [] 2025-12-04T09:50:15.8454181Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8454834Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8454917Z graph_break [] 2025-12-04T09:50:15.8455093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8455735Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8455834Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8456010Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8456104Z unimplemented [] 2025-12-04T09:50:15.8456235Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8456954Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8457040Z graph_break [] 2025-12-04T09:50:15.8457209Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8457831Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8457926Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8458103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8458190Z unimplemented [] 2025-12-04T09:50:15.8458325Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8458970Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8459100Z graph_break [] 2025-12-04T09:50:15.8459276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8459894Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8459991Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8460161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8460244Z unimplemented [] 2025-12-04T09:50:15.8460376Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8461010Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8461155Z graph_break [] 2025-12-04T09:50:15.8461335Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8461955Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8462059Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8462228Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8462310Z unimplemented [] 2025-12-04T09:50:15.8462450Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8463089Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8463176Z graph_break [] 2025-12-04T09:50:15.8463357Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8464016Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8464122Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8464298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8464380Z unimplemented [] 2025-12-04T09:50:15.8464519Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8465159Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8465242Z graph_break [] 2025-12-04T09:50:15.8465415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8466039Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8466218Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8466388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8466477Z unimplemented [] 2025-12-04T09:50:15.8466602Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8467242Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8467324Z graph_break [] 2025-12-04T09:50:15.8467490Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8468105Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8468243Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8468416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8468504Z unimplemented [] 2025-12-04T09:50:15.8468632Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8469272Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8469362Z graph_break [] 2025-12-04T09:50:15.8469535Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8470163Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8470296Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8470466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8470550Z unimplemented [] 2025-12-04T09:50:15.8470686Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8471319Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8471403Z graph_break [] 2025-12-04T09:50:15.8471574Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8472197Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8472294Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8472462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8472558Z unimplemented [] 2025-12-04T09:50:15.8472687Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8473375Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8473459Z graph_break [] 2025-12-04T09:50:15.8473626Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8474246Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8474339Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8474507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8474592Z unimplemented [] 2025-12-04T09:50:15.8474726Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8475451Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8475533Z graph_break [] 2025-12-04T09:50:15.8475702Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8476324Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8476419Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8476588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8476679Z unimplemented [] 2025-12-04T09:50:15.8476806Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8477451Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8477573Z graph_break [] 2025-12-04T09:50:15.8477747Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8478369Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8478466Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8478639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8478721Z unimplemented [] 2025-12-04T09:50:15.8478847Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8479492Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8479611Z graph_break [] 2025-12-04T09:50:15.8479788Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8480420Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8480515Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8480686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8480768Z unimplemented [] 2025-12-04T09:50:15.8480894Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8481537Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8481616Z graph_break [] 2025-12-04T09:50:15.8481782Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8482412Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8482505Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8482679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8482761Z unimplemented [] 2025-12-04T09:50:15.8482889Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8483531Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8483609Z graph_break [] 2025-12-04T09:50:15.8483777Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8484393Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8484490Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8484737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8484821Z unimplemented [] 2025-12-04T09:50:15.8484949Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8485588Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8485668Z graph_break [] 2025-12-04T09:50:15.8485839Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8486455Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8486552Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8486763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8486845Z unimplemented [] 2025-12-04T09:50:15.8486980Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8487671Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8487750Z graph_break [] 2025-12-04T09:50:15.8487923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8488540Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8488631Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8488849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8488936Z unimplemented [] 2025-12-04T09:50:15.8489066Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8489706Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8489791Z graph_break [] 2025-12-04T09:50:15.8489965Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8490580Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8490674Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8490884Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8490988Z Traceback (most recent call last): 2025-12-04T09:50:15.8491371Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8491450Z ).run(code) 2025-12-04T09:50:15.8491736Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8491821Z Searched string: 2025-12-04T09:50:15.8491898Z int32_t u0_raw; 2025-12-04T09:50:15.8492083Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8492167Z auto u0 = u0_raw; 2025-12-04T09:50:15.8492262Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8492516Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8492768Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8493067Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8493075Z 2025-12-04T09:50:15.8493230Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8493353Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8493545Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8494080Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8494203Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8494328Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8494449Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8494560Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8494992Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8495150Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8495492Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8495642Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8496023Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8496134Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8496234Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8496336Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8496344Z 2025-12-04T09:50:15.8496348Z 2025-12-04T09:50:15.8496352Z 2025-12-04T09:50:15.8496355Z 2025-12-04T09:50:15.8497021Z Wrapper code written to: /tmp/tmp8wgeup53/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cssyumg2n6nv2zh7npzbe4eykua5afxnllekqk3334hoi6mt7drz.wrapper.cpp 2025-12-04T09:50:15.8497678Z Kernel code written to: /tmp/tmp8wgeup53/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cvpxczncsdjy7i7w3xvcpnxvjq64wmxmn7aqpxszdtt3hkmbtskb.kernel.cpp 2025-12-04T09:50:15.8497881Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8497886Z 2025-12-04T09:50:15.8497889Z 2025-12-04T09:50:15.8498070Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8498585Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8498590Z 2025-12-04T09:50:15.8498806Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8498981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8499066Z unimplemented [] 2025-12-04T09:50:15.8499197Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8499755Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8499840Z graph_break [] 2025-12-04T09:50:15.8500009Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8500639Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8500734Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8500905Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8500987Z unimplemented [] 2025-12-04T09:50:15.8501117Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8501754Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8501833Z graph_break [] 2025-12-04T09:50:15.8502003Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8502706Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8502802Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8502974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8503060Z unimplemented [] 2025-12-04T09:50:15.8503210Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8503872Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8503950Z graph_break [] 2025-12-04T09:50:15.8504131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8504752Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8504884Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8505063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8505145Z unimplemented [] 2025-12-04T09:50:15.8505274Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8505917Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8505995Z graph_break [] 2025-12-04T09:50:15.8506164Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8506781Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8506914Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8507086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8507172Z unimplemented [] 2025-12-04T09:50:15.8507301Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8507947Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8508032Z graph_break [] 2025-12-04T09:50:15.8508202Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8508815Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8508915Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8509096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8509186Z unimplemented [] 2025-12-04T09:50:15.8509317Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8509960Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8510039Z graph_break [] 2025-12-04T09:50:15.8510216Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8510835Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8510928Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8511099Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8511187Z unimplemented [] 2025-12-04T09:50:15.8511326Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8512045Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8512126Z graph_break [] 2025-12-04T09:50:15.8512297Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8512914Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8513012Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8513183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8513265Z unimplemented [] 2025-12-04T09:50:15.8513395Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8514039Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8514168Z graph_break [] 2025-12-04T09:50:15.8514342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8514961Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8515056Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8515224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8515307Z unimplemented [] 2025-12-04T09:50:15.8515438Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8516080Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8516203Z graph_break [] 2025-12-04T09:50:15.8516374Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8516997Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8517093Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8517261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8517346Z unimplemented [] 2025-12-04T09:50:15.8517477Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8518118Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8518206Z graph_break [] 2025-12-04T09:50:15.8518376Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8518997Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8519097Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8519266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8519349Z unimplemented [] 2025-12-04T09:50:15.8519481Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8520121Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8520211Z graph_break [] 2025-12-04T09:50:15.8520382Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8520999Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8521179Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8521348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8521438Z unimplemented [] 2025-12-04T09:50:15.8521566Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8522209Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8522291Z graph_break [] 2025-12-04T09:50:15.8522457Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8523072Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8523211Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8523388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8523472Z unimplemented [] 2025-12-04T09:50:15.8523602Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8524237Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8524319Z graph_break [] 2025-12-04T09:50:15.8524488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8525107Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8525239Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8525406Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8525498Z unimplemented [] 2025-12-04T09:50:15.8525631Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8526271Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8526354Z graph_break [] 2025-12-04T09:50:15.8526525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8527148Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8527294Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8527463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8527551Z unimplemented [] 2025-12-04T09:50:15.8527681Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8528323Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8528405Z graph_break [] 2025-12-04T09:50:15.8528571Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8529191Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8529285Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8529452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8529536Z unimplemented [] 2025-12-04T09:50:15.8529663Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8530424Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8530507Z graph_break [] 2025-12-04T09:50:15.8530676Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8531301Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8531396Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8531564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8531649Z unimplemented [] 2025-12-04T09:50:15.8531775Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8532422Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8532545Z graph_break [] 2025-12-04T09:50:15.8532718Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8533338Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8533431Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8533602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8533690Z unimplemented [] 2025-12-04T09:50:15.8533818Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8534463Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8534587Z graph_break [] 2025-12-04T09:50:15.8534764Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8535653Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8535782Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8535956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8536040Z unimplemented [] 2025-12-04T09:50:15.8536169Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8536810Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8536890Z graph_break [] 2025-12-04T09:50:15.8537057Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8537685Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8537784Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8537957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8538040Z unimplemented [] 2025-12-04T09:50:15.8538170Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8538811Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8538891Z graph_break [] 2025-12-04T09:50:15.8539064Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8539683Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8539784Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8540099Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8540186Z unimplemented [] 2025-12-04T09:50:15.8540314Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8540955Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8541033Z graph_break [] 2025-12-04T09:50:15.8541204Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8541822Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8541920Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8542150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8542234Z unimplemented [] 2025-12-04T09:50:15.8542368Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8543009Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8543088Z graph_break [] 2025-12-04T09:50:15.8543258Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8543876Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8543969Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8544182Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8544347Z Traceback (most recent call last): 2025-12-04T09:50:15.8544727Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8544806Z ).run(code) 2025-12-04T09:50:15.8545089Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8545174Z Searched string: 2025-12-04T09:50:15.8545252Z int32_t u0_raw; 2025-12-04T09:50:15.8545437Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8545521Z auto u0 = u0_raw; 2025-12-04T09:50:15.8545617Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8545870Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8546121Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8546398Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8546406Z 2025-12-04T09:50:15.8546532Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8546659Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8546761Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8547303Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8547429Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8547558Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8547679Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8547794Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8548234Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8548393Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8548747Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8548897Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8549280Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8549390Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8549494Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8549594Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8549598Z 2025-12-04T09:50:15.8549602Z 2025-12-04T09:50:15.8549611Z 2025-12-04T09:50:15.8549615Z 2025-12-04T09:50:15.8550286Z Wrapper code written to: /tmp/tmpelvvkc93/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/ci2y3y2ehbgmq6urvxudpfsweevy453uwvzjgceaj4o5ucqbk2z5.wrapper.cpp 2025-12-04T09:50:15.8550952Z Kernel code written to: /tmp/tmpelvvkc93/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/czm3pxdnswipucfhbg3lvcx3vckszqd6ejlvqyeaz24xekalsqdm.kernel.cpp 2025-12-04T09:50:15.8551162Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8551167Z 2025-12-04T09:50:15.8551170Z 2025-12-04T09:50:15.8551347Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8551863Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8551868Z 2025-12-04T09:50:15.8552088Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8552267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8552379Z unimplemented [] 2025-12-04T09:50:15.8552525Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8553146Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8553230Z graph_break [] 2025-12-04T09:50:15.8553406Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8554035Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8554130Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8554302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8554390Z unimplemented [] 2025-12-04T09:50:15.8554520Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8555166Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8555257Z graph_break [] 2025-12-04T09:50:15.8555439Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8556069Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8561922Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8562113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8562211Z unimplemented [] 2025-12-04T09:50:15.8562364Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8563050Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8563139Z graph_break [] 2025-12-04T09:50:15.8563323Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8564059Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8564172Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8564348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8564442Z unimplemented [] 2025-12-04T09:50:15.8564577Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8565224Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8565315Z graph_break [] 2025-12-04T09:50:15.8565489Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8566120Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8566260Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8566438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8566531Z unimplemented [] 2025-12-04T09:50:15.8566666Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8567396Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8567488Z graph_break [] 2025-12-04T09:50:15.8567660Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8568288Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8568433Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8568604Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8568701Z unimplemented [] 2025-12-04T09:50:15.8568830Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8569472Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8569563Z graph_break [] 2025-12-04T09:50:15.8569732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8570364Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8570464Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8570640Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8570733Z unimplemented [] 2025-12-04T09:50:15.8570861Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8571515Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8571597Z graph_break [] 2025-12-04T09:50:15.8571769Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8572402Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8572502Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8572677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8572775Z unimplemented [] 2025-12-04T09:50:15.8572908Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8573669Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8573754Z graph_break [] 2025-12-04T09:50:15.8573926Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8574556Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8574653Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8574828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8574916Z unimplemented [] 2025-12-04T09:50:15.8575047Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8575702Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8575823Z graph_break [] 2025-12-04T09:50:15.8576002Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8576634Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8576734Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8576917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8577002Z unimplemented [] 2025-12-04T09:50:15.8577131Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8577777Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8577903Z graph_break [] 2025-12-04T09:50:15.8578074Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8578707Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8578803Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8578980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8579065Z unimplemented [] 2025-12-04T09:50:15.8579197Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8579845Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8579933Z graph_break [] 2025-12-04T09:50:15.8580117Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8580741Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8580836Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8581013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8581099Z unimplemented [] 2025-12-04T09:50:15.8581230Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8581880Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8581962Z graph_break [] 2025-12-04T09:50:15.8582141Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8582766Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8582939Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8583120Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8583208Z unimplemented [] 2025-12-04T09:50:15.8583340Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8583984Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8584065Z graph_break [] 2025-12-04T09:50:15.8584241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8584858Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8584995Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8585171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8585263Z unimplemented [] 2025-12-04T09:50:15.8585397Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8586036Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8586117Z graph_break [] 2025-12-04T09:50:15.8586292Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8586909Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8587044Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8587226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8587317Z unimplemented [] 2025-12-04T09:50:15.8587462Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8588106Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8588187Z graph_break [] 2025-12-04T09:50:15.8588362Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8588984Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8589089Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8589258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8589344Z unimplemented [] 2025-12-04T09:50:15.8589482Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8590132Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8590216Z graph_break [] 2025-12-04T09:50:15.8590391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8591012Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8591117Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8591288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8591377Z unimplemented [] 2025-12-04T09:50:15.8591516Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8592214Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8592421Z graph_break [] 2025-12-04T09:50:15.8592637Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8593409Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8593539Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8593750Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8593856Z unimplemented [] 2025-12-04T09:50:15.8594028Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8594796Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8594929Z graph_break [] 2025-12-04T09:50:15.8595105Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8595724Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8595827Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8596003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8596088Z unimplemented [] 2025-12-04T09:50:15.8596228Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8596867Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8596994Z graph_break [] 2025-12-04T09:50:15.8597164Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8597791Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8597895Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8598067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8598160Z unimplemented [] 2025-12-04T09:50:15.8598290Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8598927Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8599016Z graph_break [] 2025-12-04T09:50:15.8599188Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8599807Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8599914Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8600085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8600177Z unimplemented [] 2025-12-04T09:50:15.8600306Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8600946Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8601039Z graph_break [] 2025-12-04T09:50:15.8601211Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8601839Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8601942Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8602250Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8602366Z unimplemented [] 2025-12-04T09:50:15.8602529Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8603333Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8603443Z graph_break [] 2025-12-04T09:50:15.8603656Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8604346Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8604445Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8604661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8604753Z unimplemented [] 2025-12-04T09:50:15.8604890Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8605532Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8605622Z graph_break [] 2025-12-04T09:50:15.8605792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8606417Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8606512Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8606681Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8606833Z unimplemented [] 2025-12-04T09:50:15.8606966Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8607675Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8607759Z graph_break [] 2025-12-04T09:50:15.8607931Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8608554Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8608650Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8608821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8608913Z unimplemented [] 2025-12-04T09:50:15.8609047Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8609700Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8609788Z graph_break [] 2025-12-04T09:50:15.8609958Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8610584Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8610682Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8610895Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8611008Z Traceback (most recent call last): 2025-12-04T09:50:15.8611383Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8611472Z ).run(code) 2025-12-04T09:50:15.8611762Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8611848Z Searched string: 2025-12-04T09:50:15.8612022Z int32_t u0_raw; 2025-12-04T09:50:15.8612213Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8612300Z auto u0 = u0_raw; 2025-12-04T09:50:15.8612425Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8612714Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8612971Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8613248Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8613253Z 2025-12-04T09:50:15.8613387Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8613522Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8613671Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8614224Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8614352Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8614482Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8614614Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8614729Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8615165Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8615333Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8615603Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8615798Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8616195Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8616302Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8616414Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8616515Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8616519Z 2025-12-04T09:50:15.8616523Z 2025-12-04T09:50:15.8616527Z 2025-12-04T09:50:15.8616531Z 2025-12-04T09:50:15.8617219Z Wrapper code written to: /tmp/tmpmnablfyn/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/ccdx57szd4hcekqnkyjtrsxg3fyrjyzdc6xhkrzd5uxxzt7cpgnz.wrapper.cpp 2025-12-04T09:50:15.8617884Z Kernel code written to: /tmp/tmpmnablfyn/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cfneaoqujuwu67slsls2zwmniprcvyggjanfxxsts547ol7itigh.kernel.cpp 2025-12-04T09:50:15.8618047Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8618062Z 2025-12-04T09:50:15.8618067Z 2025-12-04T09:50:15.8618247Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8618761Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8618766Z 2025-12-04T09:50:15.8618995Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8619172Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8619266Z unimplemented [] 2025-12-04T09:50:15.8619398Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8619958Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8620052Z graph_break [] 2025-12-04T09:50:15.8620226Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8620935Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8621043Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8621215Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8621313Z unimplemented [] 2025-12-04T09:50:15.8621449Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8622096Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8622187Z graph_break [] 2025-12-04T09:50:15.8622359Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8622987Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8623128Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8623306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8623401Z unimplemented [] 2025-12-04T09:50:15.8623536Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8624177Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8624271Z graph_break [] 2025-12-04T09:50:15.8624441Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8625068Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8625207Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8625383Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8625477Z unimplemented [] 2025-12-04T09:50:15.8625613Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8626258Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8626348Z graph_break [] 2025-12-04T09:50:15.8626519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8627145Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8627241Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8627415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8627513Z unimplemented [] 2025-12-04T09:50:15.8627643Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8628299Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8628380Z graph_break [] 2025-12-04T09:50:15.8628551Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8629176Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8629271Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8629443Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8629533Z unimplemented [] 2025-12-04T09:50:15.8629668Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8630392Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8630478Z graph_break [] 2025-12-04T09:50:15.8630648Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8631280Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8631378Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8631548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8631640Z unimplemented [] 2025-12-04T09:50:15.8631775Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8632428Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8632553Z graph_break [] 2025-12-04T09:50:15.8632731Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8633414Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8633512Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8633689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8633774Z unimplemented [] 2025-12-04T09:50:15.8633904Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8634553Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8634682Z graph_break [] 2025-12-04T09:50:15.8634853Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8635796Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8635899Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8636079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8636165Z unimplemented [] 2025-12-04T09:50:15.8636298Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8636939Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8637022Z graph_break [] 2025-12-04T09:50:15.8637195Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8637829Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8637927Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8638108Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8638193Z unimplemented [] 2025-12-04T09:50:15.8638323Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8638973Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8639057Z graph_break [] 2025-12-04T09:50:15.8639235Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8639855Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8639954Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8640271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8640363Z unimplemented [] 2025-12-04T09:50:15.8640496Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8641144Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8641225Z graph_break [] 2025-12-04T09:50:15.8641404Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8642025Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8642208Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8642388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8642480Z unimplemented [] 2025-12-04T09:50:15.8642618Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8643256Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8643338Z graph_break [] 2025-12-04T09:50:15.8643523Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8644143Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8644240Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8644483Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8644576Z unimplemented [] 2025-12-04T09:50:15.8644714Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8645363Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8645447Z graph_break [] 2025-12-04T09:50:15.8645631Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8646250Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8646353Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8646525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8646616Z unimplemented [] 2025-12-04T09:50:15.8646754Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8647470Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8647555Z graph_break [] 2025-12-04T09:50:15.8647738Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8648361Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8648465Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8648639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8648726Z unimplemented [] 2025-12-04T09:50:15.8648865Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8649513Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8649684Z graph_break [] 2025-12-04T09:50:15.8649867Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8650488Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8650591Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8650763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8650849Z unimplemented [] 2025-12-04T09:50:15.8650987Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8651633Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8651769Z graph_break [] 2025-12-04T09:50:15.8651941Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8652623Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8652729Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8652902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8652989Z unimplemented [] 2025-12-04T09:50:15.8653127Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8653773Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8653899Z graph_break [] 2025-12-04T09:50:15.8654071Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8654712Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8654809Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8654987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8655075Z unimplemented [] 2025-12-04T09:50:15.8655208Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8655860Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8655942Z graph_break [] 2025-12-04T09:50:15.8656116Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8656748Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8656854Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8657031Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8657120Z unimplemented [] 2025-12-04T09:50:15.8657250Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8657900Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8657990Z graph_break [] 2025-12-04T09:50:15.8658167Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8658791Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8658893Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8659155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8659244Z unimplemented [] 2025-12-04T09:50:15.8659376Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8660027Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8660110Z graph_break [] 2025-12-04T09:50:15.8660293Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8660912Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8661015Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8661193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8661319Z unimplemented [] 2025-12-04T09:50:15.8661466Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8662107Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8662189Z graph_break [] 2025-12-04T09:50:15.8662367Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8662991Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8663089Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8663267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8663394Z unimplemented [] 2025-12-04T09:50:15.8663533Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8664185Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8664266Z graph_break [] 2025-12-04T09:50:15.8664448Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8665072Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8665172Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8665344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8665430Z unimplemented [] 2025-12-04T09:50:15.8665569Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8666214Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8666304Z graph_break [] 2025-12-04T09:50:15.8666485Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8667109Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8667211Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8667382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8667467Z unimplemented [] 2025-12-04T09:50:15.8667601Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8668242Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8668334Z graph_break [] 2025-12-04T09:50:15.8668582Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8669203Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8669308Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8669478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8669563Z unimplemented [] 2025-12-04T09:50:15.8669698Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8670346Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8670440Z graph_break [] 2025-12-04T09:50:15.8670650Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8671277Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8671383Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8671554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8671641Z unimplemented [] 2025-12-04T09:50:15.8671776Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8672420Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8672508Z graph_break [] 2025-12-04T09:50:15.8672678Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8673341Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8673451Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8673622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8673715Z unimplemented [] 2025-12-04T09:50:15.8673845Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8674490Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8674581Z graph_break [] 2025-12-04T09:50:15.8674751Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8675373Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8675480Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8675656Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8675749Z unimplemented [] 2025-12-04T09:50:15.8675879Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8676523Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8676612Z graph_break [] 2025-12-04T09:50:15.8676788Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8677414Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8677513Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8677687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8677780Z unimplemented [] 2025-12-04T09:50:15.8678012Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8678659Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8678747Z graph_break [] 2025-12-04T09:50:15.8678917Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8679544Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8679641Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8679853Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8680007Z Traceback (most recent call last): 2025-12-04T09:50:15.8680387Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8680470Z ).run(code) 2025-12-04T09:50:15.8680764Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8680848Z Searched string: 2025-12-04T09:50:15.8680935Z int32_t u0_raw; 2025-12-04T09:50:15.8681123Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8681208Z auto u0 = u0_raw; 2025-12-04T09:50:15.8681311Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8681569Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8681821Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8682150Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8682157Z 2025-12-04T09:50:15.8682284Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8682426Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8682532Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8683075Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8683209Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8683336Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8683465Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8683584Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8684021Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8684193Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8684469Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8684619Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8685014Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8685119Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8685229Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8685331Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8685335Z 2025-12-04T09:50:15.8685339Z 2025-12-04T09:50:15.8685343Z 2025-12-04T09:50:15.8685347Z 2025-12-04T09:50:15.8686035Z Wrapper code written to: /tmp/tmpkylini8i/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cwczcj7eicdl2qcsmcnmshfpeahnrahk6ra7vqxgqmlrtddgoesg.wrapper.cpp 2025-12-04T09:50:15.8686696Z Kernel code written to: /tmp/tmpkylini8i/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cmn4evgqb3ylk4lnpi6wpfikeou73h5xiyv3o7sehn7t2g272d6w.kernel.cpp 2025-12-04T09:50:15.8686943Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8686948Z 2025-12-04T09:50:15.8686952Z 2025-12-04T09:50:15.8687141Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8687727Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8687732Z 2025-12-04T09:50:15.8687958Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8688135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8688220Z unimplemented [] 2025-12-04T09:50:15.8688358Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8688918Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8689050Z graph_break [] 2025-12-04T09:50:15.8689231Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8689860Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8689964Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8690137Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8690223Z unimplemented [] 2025-12-04T09:50:15.8690362Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8691009Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8691140Z graph_break [] 2025-12-04T09:50:15.8691313Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8691944Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8692045Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8692218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8692303Z unimplemented [] 2025-12-04T09:50:15.8692437Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8693082Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8693168Z graph_break [] 2025-12-04T09:50:15.8693342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8693977Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8694082Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8694255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8694348Z unimplemented [] 2025-12-04T09:50:15.8694481Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8695127Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8695216Z graph_break [] 2025-12-04T09:50:15.8695386Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8696008Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8696114Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8696365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8696461Z unimplemented [] 2025-12-04T09:50:15.8696597Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8697238Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8697327Z graph_break [] 2025-12-04T09:50:15.8697496Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8698126Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8698263Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8698434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8698529Z unimplemented [] 2025-12-04T09:50:15.8698659Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8699304Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8699394Z graph_break [] 2025-12-04T09:50:15.8699564Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8700190Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8700286Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8700504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8700597Z unimplemented [] 2025-12-04T09:50:15.8700726Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8701375Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8701464Z graph_break [] 2025-12-04T09:50:15.8701633Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8702263Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8702360Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8702530Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8702624Z unimplemented [] 2025-12-04T09:50:15.8702759Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8703420Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8703503Z graph_break [] 2025-12-04T09:50:15.8703673Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8704303Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8704400Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8704574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8704666Z unimplemented [] 2025-12-04T09:50:15.8704797Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8705451Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8705537Z graph_break [] 2025-12-04T09:50:15.8705792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8706424Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8706522Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8706693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8706786Z unimplemented [] 2025-12-04T09:50:15.8706920Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8707572Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8707697Z graph_break [] 2025-12-04T09:50:15.8707869Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8708508Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8708608Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8708785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8708870Z unimplemented [] 2025-12-04T09:50:15.8709004Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8709656Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8709739Z graph_break [] 2025-12-04T09:50:15.8709951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8710591Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8710688Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8710865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8710953Z unimplemented [] 2025-12-04T09:50:15.8711083Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8711736Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8711818Z graph_break [] 2025-12-04T09:50:15.8711994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8712618Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8712721Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8712899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8712984Z unimplemented [] 2025-12-04T09:50:15.8713114Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8713766Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8713848Z graph_break [] 2025-12-04T09:50:15.8714025Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8714645Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8714745Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8714925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8715114Z unimplemented [] 2025-12-04T09:50:15.8715247Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8715903Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8715984Z graph_break [] 2025-12-04T09:50:15.8716161Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8716780Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8716879Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8717059Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8717183Z unimplemented [] 2025-12-04T09:50:15.8717319Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8717969Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8718051Z graph_break [] 2025-12-04T09:50:15.8718231Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8718850Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8718946Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8719124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8719256Z unimplemented [] 2025-12-04T09:50:15.8719399Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8720046Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8720131Z graph_break [] 2025-12-04T09:50:15.8720312Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8720932Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8721035Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8721208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8721293Z unimplemented [] 2025-12-04T09:50:15.8721432Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8722076Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8722164Z graph_break [] 2025-12-04T09:50:15.8722342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8722966Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8723068Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8723239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8723328Z unimplemented [] 2025-12-04T09:50:15.8723464Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8724111Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8724198Z graph_break [] 2025-12-04T09:50:15.8724456Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8725079Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8725187Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8725359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8725444Z unimplemented [] 2025-12-04T09:50:15.8725584Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8726225Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8726317Z graph_break [] 2025-12-04T09:50:15.8726490Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8727158Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8727330Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8727504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8727589Z unimplemented [] 2025-12-04T09:50:15.8727727Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8728371Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8728459Z graph_break [] 2025-12-04T09:50:15.8728631Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8729297Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8729410Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8729582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8729668Z unimplemented [] 2025-12-04T09:50:15.8729805Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8730453Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8730542Z graph_break [] 2025-12-04T09:50:15.8730714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8731335Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8731447Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8731623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8731719Z unimplemented [] 2025-12-04T09:50:15.8731852Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8732497Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8732588Z graph_break [] 2025-12-04T09:50:15.8732759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8733379Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8733484Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8733659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8733756Z unimplemented [] 2025-12-04T09:50:15.8733965Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8734606Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8734698Z graph_break [] 2025-12-04T09:50:15.8734867Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8735770Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8735870Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8736042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8736139Z unimplemented [] 2025-12-04T09:50:15.8736345Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8736989Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8737079Z graph_break [] 2025-12-04T09:50:15.8737252Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8737879Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8737977Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8738150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8738245Z unimplemented [] 2025-12-04T09:50:15.8738376Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8739092Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8739180Z graph_break [] 2025-12-04T09:50:15.8739350Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8739975Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8740071Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8740241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8740333Z unimplemented [] 2025-12-04T09:50:15.8740463Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8741113Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8741200Z graph_break [] 2025-12-04T09:50:15.8741375Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8742000Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8742096Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8742266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8742358Z unimplemented [] 2025-12-04T09:50:15.8742488Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8743137Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8743220Z graph_break [] 2025-12-04T09:50:15.8743393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8744150Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8744249Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8744428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8744513Z unimplemented [] 2025-12-04T09:50:15.8744644Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8745293Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8745374Z graph_break [] 2025-12-04T09:50:15.8745545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8746222Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8746318Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8746497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8746581Z unimplemented [] 2025-12-04T09:50:15.8746711Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8747364Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8747446Z graph_break [] 2025-12-04T09:50:15.8747616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8748244Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8748436Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8748620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8748707Z unimplemented [] 2025-12-04T09:50:15.8748840Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8749490Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8749572Z graph_break [] 2025-12-04T09:50:15.8749748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8750370Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8750469Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8750649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8750733Z unimplemented [] 2025-12-04T09:50:15.8750867Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8751518Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8751600Z graph_break [] 2025-12-04T09:50:15.8751777Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8752398Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8752495Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8752681Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8752775Z unimplemented [] 2025-12-04T09:50:15.8757498Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8758284Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8758380Z graph_break [] 2025-12-04T09:50:15.8758564Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8759201Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8759302Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8759514Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8759634Z Traceback (most recent call last): 2025-12-04T09:50:15.8760012Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8760134Z ).run(code) 2025-12-04T09:50:15.8760437Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8760524Z Searched string: 2025-12-04T09:50:15.8760613Z int32_t u0_raw; 2025-12-04T09:50:15.8760805Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8760892Z auto u0 = u0_raw; 2025-12-04T09:50:15.8760999Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8761262Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8761517Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8761803Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8761852Z 2025-12-04T09:50:15.8761986Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8762124Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8762240Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8762780Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8762917Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8763047Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8763179Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8763294Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8763735Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8763903Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8764181Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8764333Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8764733Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8764843Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8764957Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8765062Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8765067Z 2025-12-04T09:50:15.8765071Z 2025-12-04T09:50:15.8765075Z 2025-12-04T09:50:15.8765079Z 2025-12-04T09:50:15.8765735Z Wrapper code written to: /tmp/tmpc0i_15m8/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cen6wyrz2ysaa3zsesy4as76nzvmcin7ihspqhylnz7rclhppx2z.wrapper.cpp 2025-12-04T09:50:15.8766378Z Kernel code written to: /tmp/tmpc0i_15m8/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cey7dmpyvk26nj2nubouhfg6rccew7qvuurlrdrtb5vozh5chmfa.kernel.cpp 2025-12-04T09:50:15.8766547Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8766551Z 2025-12-04T09:50:15.8766555Z 2025-12-04T09:50:15.8766823Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8767427Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8767433Z 2025-12-04T09:50:15.8767661Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8767842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8767930Z unimplemented [] 2025-12-04T09:50:15.8768072Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8768638Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8768769Z graph_break [] 2025-12-04T09:50:15.8768953Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8769595Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8769706Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8769882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8769972Z unimplemented [] 2025-12-04T09:50:15.8770113Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8770762Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8770856Z graph_break [] 2025-12-04T09:50:15.8771072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8771703Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8771809Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8771982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8772069Z unimplemented [] 2025-12-04T09:50:15.8772209Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8772850Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8772939Z graph_break [] 2025-12-04T09:50:15.8773110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8773736Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8773842Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8774019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8774107Z unimplemented [] 2025-12-04T09:50:15.8774245Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8774892Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8774980Z graph_break [] 2025-12-04T09:50:15.8775152Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8775772Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8775884Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8776058Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8776229Z unimplemented [] 2025-12-04T09:50:15.8776365Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8777007Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8777099Z graph_break [] 2025-12-04T09:50:15.8777272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8777893Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8777999Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8778173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8778309Z unimplemented [] 2025-12-04T09:50:15.8778441Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8779093Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8779183Z graph_break [] 2025-12-04T09:50:15.8779355Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8779985Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8780082Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8780255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8780389Z unimplemented [] 2025-12-04T09:50:15.8780522Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8781171Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8781263Z graph_break [] 2025-12-04T09:50:15.8781435Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8782062Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8782160Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8782337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8782430Z unimplemented [] 2025-12-04T09:50:15.8782562Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8783212Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8783311Z graph_break [] 2025-12-04T09:50:15.8783486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8784114Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8784213Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8784387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8784482Z unimplemented [] 2025-12-04T09:50:15.8784615Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8785267Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8785356Z graph_break [] 2025-12-04T09:50:15.8785527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8786233Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8786335Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8786508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8786602Z unimplemented [] 2025-12-04T09:50:15.8786733Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8787379Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8787465Z graph_break [] 2025-12-04T09:50:15.8787638Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8788341Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8788440Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8788624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8788710Z unimplemented [] 2025-12-04T09:50:15.8788844Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8789495Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8789578Z graph_break [] 2025-12-04T09:50:15.8789751Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8790422Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8790527Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8790708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8790795Z unimplemented [] 2025-12-04T09:50:15.8790927Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8791574Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8791661Z graph_break [] 2025-12-04T09:50:15.8791834Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8792460Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8792564Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8792746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8792833Z unimplemented [] 2025-12-04T09:50:15.8792965Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8793619Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8793702Z graph_break [] 2025-12-04T09:50:15.8793881Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8794500Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8794600Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8794778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8794867Z unimplemented [] 2025-12-04T09:50:15.8795075Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8795722Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8795804Z graph_break [] 2025-12-04T09:50:15.8795990Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8796613Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8796712Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8796896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8796984Z unimplemented [] 2025-12-04T09:50:15.8797178Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8797846Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8797932Z graph_break [] 2025-12-04T09:50:15.8798116Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8798740Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8798840Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8799017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8799105Z unimplemented [] 2025-12-04T09:50:15.8799243Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8799927Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8800018Z graph_break [] 2025-12-04T09:50:15.8800206Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8800827Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8800930Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8801106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8801194Z unimplemented [] 2025-12-04T09:50:15.8801330Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8801975Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8802061Z graph_break [] 2025-12-04T09:50:15.8802248Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8802869Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8802974Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8803151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8803239Z unimplemented [] 2025-12-04T09:50:15.8803377Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8804017Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8804101Z graph_break [] 2025-12-04T09:50:15.8804279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8804980Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8805088Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8805260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8805346Z unimplemented [] 2025-12-04T09:50:15.8805484Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8806124Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8806217Z graph_break [] 2025-12-04T09:50:15.8806391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8807015Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8807174Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8807396Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8807487Z unimplemented [] 2025-12-04T09:50:15.8807635Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8808282Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8808373Z graph_break [] 2025-12-04T09:50:15.8808549Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8809169Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8809320Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8809504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8809592Z unimplemented [] 2025-12-04T09:50:15.8809737Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8810384Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8810473Z graph_break [] 2025-12-04T09:50:15.8810647Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8811274Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8811385Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8811565Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8811663Z unimplemented [] 2025-12-04T09:50:15.8811800Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8812446Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8812543Z graph_break [] 2025-12-04T09:50:15.8812721Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8813342Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8813448Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8813621Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8813723Z unimplemented [] 2025-12-04T09:50:15.8813858Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8814585Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8814678Z graph_break [] 2025-12-04T09:50:15.8814857Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8815487Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8815584Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8815758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8815850Z unimplemented [] 2025-12-04T09:50:15.8815986Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8816686Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8816777Z graph_break [] 2025-12-04T09:50:15.8816949Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8817578Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8817675Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8817848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8817943Z unimplemented [] 2025-12-04T09:50:15.8818076Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8818721Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8818856Z graph_break [] 2025-12-04T09:50:15.8819036Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8819671Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8819767Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8819939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8820035Z unimplemented [] 2025-12-04T09:50:15.8820168Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8820820Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8820908Z graph_break [] 2025-12-04T09:50:15.8821082Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8821715Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8821813Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8821986Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8822078Z unimplemented [] 2025-12-04T09:50:15.8822210Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8822912Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8822993Z graph_break [] 2025-12-04T09:50:15.8823167Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8823905Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8824006Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8824182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8824267Z unimplemented [] 2025-12-04T09:50:15.8824402Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8825044Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8825125Z graph_break [] 2025-12-04T09:50:15.8825298Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8825923Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8826064Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8826256Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8826345Z unimplemented [] 2025-12-04T09:50:15.8826478Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8827139Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8827223Z graph_break [] 2025-12-04T09:50:15.8827396Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8828022Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8828163Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8828350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8828443Z unimplemented [] 2025-12-04T09:50:15.8828586Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8829240Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8829321Z graph_break [] 2025-12-04T09:50:15.8829500Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8830128Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8830224Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8830409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8830498Z unimplemented [] 2025-12-04T09:50:15.8830632Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8831294Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8831383Z graph_break [] 2025-12-04T09:50:15.8831572Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8832194Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8832292Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8832475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8832561Z unimplemented [] 2025-12-04T09:50:15.8832701Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8833475Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8833572Z graph_break [] 2025-12-04T09:50:15.8833755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8834379Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8834477Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8834659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8834746Z unimplemented [] 2025-12-04T09:50:15.8834885Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8835828Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8835999Z graph_break [] 2025-12-04T09:50:15.8836183Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8836806Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8836905Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8837086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8837174Z unimplemented [] 2025-12-04T09:50:15.8837321Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8837968Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8838116Z graph_break [] 2025-12-04T09:50:15.8838302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8838933Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8839035Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8839248Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8839357Z Traceback (most recent call last): 2025-12-04T09:50:15.8839747Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8839827Z ).run(code) 2025-12-04T09:50:15.8840122Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8840217Z Searched string: 2025-12-04T09:50:15.8840302Z int32_t u0_raw; 2025-12-04T09:50:15.8840499Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8840595Z auto u0 = u0_raw; 2025-12-04T09:50:15.8840698Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8840968Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8841226Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8841502Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8841515Z 2025-12-04T09:50:15.8841643Z const int64_t int_array_4[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8841768Z const int64_t int_array_5[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8841885Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8842426Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_4, int_array_5, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8842562Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8842816Z const int64_t int_array_6[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8842945Z const int64_t int_array_7[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8843072Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8843509Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_6, int_array_7, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8843675Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8843960Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8844111Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8844497Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8844657Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8844762Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8844881Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8844890Z 2025-12-04T09:50:15.8844894Z 2025-12-04T09:50:15.8844898Z 2025-12-04T09:50:15.8844902Z 2025-12-04T09:50:15.8845579Z Wrapper code written to: /tmp/tmpn2z2gjfp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/clew3iphn6yh3tmx5hxgcs3rcauyxhzlzdsa4ftdm6pr7tqy4oav.wrapper.cpp 2025-12-04T09:50:15.8846249Z Kernel code written to: /tmp/tmpn2z2gjfp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cynr3jtbwdo5ke3poekmiwdi7jlwki3jvricsnupij5ddug7jorz.kernel.cpp 2025-12-04T09:50:15.8846410Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8846415Z 2025-12-04T09:50:15.8846418Z 2025-12-04T09:50:15.8846602Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8847165Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8847173Z 2025-12-04T09:50:15.8847478Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8847662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8847749Z unimplemented [] 2025-12-04T09:50:15.8847881Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8848446Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8848526Z graph_break [] 2025-12-04T09:50:15.8848715Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8849345Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8849454Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8849627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8849715Z unimplemented [] 2025-12-04T09:50:15.8849854Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8850502Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8850590Z graph_break [] 2025-12-04T09:50:15.8850761Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8851381Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8851488Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8851663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8851751Z unimplemented [] 2025-12-04T09:50:15.8851894Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8852624Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8852718Z graph_break [] 2025-12-04T09:50:15.8852892Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8853522Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8853634Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8853805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8853906Z unimplemented [] 2025-12-04T09:50:15.8854078Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8854725Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8854816Z graph_break [] 2025-12-04T09:50:15.8854994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8855625Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8855723Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8855898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8855994Z unimplemented [] 2025-12-04T09:50:15.8856129Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8856813Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8856915Z graph_break [] 2025-12-04T09:50:15.8857087Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8857713Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8857812Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8857985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8858075Z unimplemented [] 2025-12-04T09:50:15.8858205Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8858847Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8858943Z graph_break [] 2025-12-04T09:50:15.8859118Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8859746Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8859845Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8860016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8860115Z unimplemented [] 2025-12-04T09:50:15.8860249Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8860905Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8860995Z graph_break [] 2025-12-04T09:50:15.8861166Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8861933Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8862032Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8862205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8862305Z unimplemented [] 2025-12-04T09:50:15.8862440Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8863098Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8863179Z graph_break [] 2025-12-04T09:50:15.8863350Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8863981Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8864122Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8864301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8864390Z unimplemented [] 2025-12-04T09:50:15.8864520Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8865170Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8865252Z graph_break [] 2025-12-04T09:50:15.8865422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8866058Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8866203Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8866390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8866476Z unimplemented [] 2025-12-04T09:50:15.8866607Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8867254Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8867336Z graph_break [] 2025-12-04T09:50:15.8867506Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8868134Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8868235Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8868418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8868508Z unimplemented [] 2025-12-04T09:50:15.8868648Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8869296Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8869379Z graph_break [] 2025-12-04T09:50:15.8869560Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8870181Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8870280Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8870458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8870550Z unimplemented [] 2025-12-04T09:50:15.8870684Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8871416Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8871502Z graph_break [] 2025-12-04T09:50:15.8871679Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8872303Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8872401Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8872584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8872671Z unimplemented [] 2025-12-04T09:50:15.8872805Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8873515Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8873641Z graph_break [] 2025-12-04T09:50:15.8873818Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8874441Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8874541Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8874722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8874806Z unimplemented [] 2025-12-04T09:50:15.8874943Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8875582Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8875705Z graph_break [] 2025-12-04T09:50:15.8875887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8876508Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8876610Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8876783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8876870Z unimplemented [] 2025-12-04T09:50:15.8877013Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8877661Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8877749Z graph_break [] 2025-12-04T09:50:15.8877930Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8878560Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8878669Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8878846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8878938Z unimplemented [] 2025-12-04T09:50:15.8879083Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8879732Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8879818Z graph_break [] 2025-12-04T09:50:15.8879997Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8880701Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8880808Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8880982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8881068Z unimplemented [] 2025-12-04T09:50:15.8881204Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8881853Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8881942Z graph_break [] 2025-12-04T09:50:15.8882116Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8882735Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8882883Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8883089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8883189Z unimplemented [] 2025-12-04T09:50:15.8883344Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8883984Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8884074Z graph_break [] 2025-12-04T09:50:15.8884245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8884869Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8885013Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8885189Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8885275Z unimplemented [] 2025-12-04T09:50:15.8885419Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8886065Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8886151Z graph_break [] 2025-12-04T09:50:15.8886324Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8886943Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8887051Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8887221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8887367Z unimplemented [] 2025-12-04T09:50:15.8887504Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8888157Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8888248Z graph_break [] 2025-12-04T09:50:15.8888421Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8889039Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8889141Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8889311Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8889403Z unimplemented [] 2025-12-04T09:50:15.8889536Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8890266Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8890356Z graph_break [] 2025-12-04T09:50:15.8890526Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8891155Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8891254Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8891423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8891513Z unimplemented [] 2025-12-04T09:50:15.8891648Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8892293Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8892424Z graph_break [] 2025-12-04T09:50:15.8892602Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8897717Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8897925Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8898198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8898336Z unimplemented [] 2025-12-04T09:50:15.8898530Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8899211Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8899399Z graph_break [] 2025-12-04T09:50:15.8899577Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8900219Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8900323Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8900529Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8900618Z unimplemented [] 2025-12-04T09:50:15.8900768Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8901410Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8901495Z graph_break [] 2025-12-04T09:50:15.8901676Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8902307Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8902412Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8902590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8902675Z unimplemented [] 2025-12-04T09:50:15.8902813Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8903453Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8903537Z graph_break [] 2025-12-04T09:50:15.8903717Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8904334Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8904445Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8904665Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8904753Z unimplemented [] 2025-12-04T09:50:15.8904893Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8905535Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8905622Z graph_break [] 2025-12-04T09:50:15.8905795Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8906416Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8906560Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8906731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8906821Z unimplemented [] 2025-12-04T09:50:15.8906957Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8907670Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8907762Z graph_break [] 2025-12-04T09:50:15.8907933Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8908551Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8908654Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8908869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8908962Z unimplemented [] 2025-12-04T09:50:15.8909099Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8909737Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8909827Z graph_break [] 2025-12-04T09:50:15.8910001Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8910625Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8910728Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8910899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8910992Z unimplemented [] 2025-12-04T09:50:15.8911123Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8911767Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8911854Z graph_break [] 2025-12-04T09:50:15.8912026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8912648Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8912752Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8912922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8913012Z unimplemented [] 2025-12-04T09:50:15.8913145Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8913782Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8913871Z graph_break [] 2025-12-04T09:50:15.8914094Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8914726Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8914822Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8914995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8915090Z unimplemented [] 2025-12-04T09:50:15.8915220Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8915855Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8915990Z graph_break [] 2025-12-04T09:50:15.8916164Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8916799Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8916895Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8917128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8917222Z unimplemented [] 2025-12-04T09:50:15.8917353Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8917994Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8918082Z graph_break [] 2025-12-04T09:50:15.8918297Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8918939Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8919035Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8919208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8919307Z unimplemented [] 2025-12-04T09:50:15.8919441Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8920092Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8920175Z graph_break [] 2025-12-04T09:50:15.8920352Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8920987Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8921089Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8921261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8921356Z unimplemented [] 2025-12-04T09:50:15.8921492Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8922143Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8922226Z graph_break [] 2025-12-04T09:50:15.8922399Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8923029Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8923132Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8923313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8923443Z unimplemented [] 2025-12-04T09:50:15.8923578Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8924237Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8924319Z graph_break [] 2025-12-04T09:50:15.8924492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8925120Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8925215Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8925395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8925522Z unimplemented [] 2025-12-04T09:50:15.8925656Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8926315Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8926397Z graph_break [] 2025-12-04T09:50:15.8926611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8927308Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8927406Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8927630Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.8927781Z Traceback (most recent call last): 2025-12-04T09:50:15.8928163Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.8928247Z ).run(code) 2025-12-04T09:50:15.8928538Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.8928622Z Searched string: 2025-12-04T09:50:15.8928710Z int32_t u0_raw; 2025-12-04T09:50:15.8928901Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.8928989Z auto u0 = u0_raw; 2025-12-04T09:50:15.8929086Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.8929345Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8929600Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8929875Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.8929886Z 2025-12-04T09:50:15.8930019Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8930147Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8930254Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.8930802Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.8930930Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.8931055Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8931185Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.8931300Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.8931741Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.8931901Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.8932174Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.8932383Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.8932774Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.8932890Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.8932993Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.8933094Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.8933098Z 2025-12-04T09:50:15.8933102Z 2025-12-04T09:50:15.8933106Z 2025-12-04T09:50:15.8933109Z 2025-12-04T09:50:15.8933787Z Wrapper code written to: /tmp/tmp2prvqs7d/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c4g2b4ohpfuu43y73weckxfatc2nb7y7lk74z6qou2k4bk7uomxo.wrapper.cpp 2025-12-04T09:50:15.8934453Z Kernel code written to: /tmp/tmp2prvqs7d/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cbgblnxqluncaedgthvheljmtzs5tfg2cy3awug4lgqjepjrp7am.kernel.cpp 2025-12-04T09:50:15.8934687Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.8934694Z 2025-12-04T09:50:15.8934698Z 2025-12-04T09:50:15.8934880Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.8935731Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.8935744Z 2025-12-04T09:50:15.8935971Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.8936150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8936242Z unimplemented [] 2025-12-04T09:50:15.8936376Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8936942Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8937098Z graph_break [] 2025-12-04T09:50:15.8937278Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8937913Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8938017Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8938193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8938287Z unimplemented [] 2025-12-04T09:50:15.8938420Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8939058Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8939149Z graph_break [] 2025-12-04T09:50:15.8939322Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8939952Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8940049Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8940224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8940319Z unimplemented [] 2025-12-04T09:50:15.8940450Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8941091Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8941174Z graph_break [] 2025-12-04T09:50:15.8941348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8941986Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8942149Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8942348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8942445Z unimplemented [] 2025-12-04T09:50:15.8942596Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8943247Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8943333Z graph_break [] 2025-12-04T09:50:15.8943506Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8944132Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8944289Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8944471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8944557Z unimplemented [] 2025-12-04T09:50:15.8944692Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8945378Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8945463Z graph_break [] 2025-12-04T09:50:15.8945634Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8946263Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8946396Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8946576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8946666Z unimplemented [] 2025-12-04T09:50:15.8946800Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8947442Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8947526Z graph_break [] 2025-12-04T09:50:15.8947698Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8948326Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8948424Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8948607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8948700Z unimplemented [] 2025-12-04T09:50:15.8948836Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8949495Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8949578Z graph_break [] 2025-12-04T09:50:15.8949756Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8950377Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8950473Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8950647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8950733Z unimplemented [] 2025-12-04T09:50:15.8950868Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8951560Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8951644Z graph_break [] 2025-12-04T09:50:15.8951823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8952450Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8952548Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8952730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8957439Z unimplemented [] 2025-12-04T09:50:15.8957582Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8958243Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8958406Z graph_break [] 2025-12-04T09:50:15.8958592Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8959235Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8959383Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8959572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8959661Z unimplemented [] 2025-12-04T09:50:15.8959798Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8960447Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8960572Z graph_break [] 2025-12-04T09:50:15.8960751Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8961390Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8961489Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8961672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8961760Z unimplemented [] 2025-12-04T09:50:15.8961900Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8962549Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8962633Z graph_break [] 2025-12-04T09:50:15.8962808Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8963444Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8963544Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8963725Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8963812Z unimplemented [] 2025-12-04T09:50:15.8963949Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8964594Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8964679Z graph_break [] 2025-12-04T09:50:15.8964857Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8965481Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8965580Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8965808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8965896Z unimplemented [] 2025-12-04T09:50:15.8966028Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8966681Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8966765Z graph_break [] 2025-12-04T09:50:15.8966947Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8967639Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8967742Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8967969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8968057Z unimplemented [] 2025-12-04T09:50:15.8968191Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8968887Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8968971Z graph_break [] 2025-12-04T09:50:15.8969150Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8969777Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8969877Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8970100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8970190Z unimplemented [] 2025-12-04T09:50:15.8970327Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8970970Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8971055Z graph_break [] 2025-12-04T09:50:15.8971238Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8971859Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8971955Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8972136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8972221Z unimplemented [] 2025-12-04T09:50:15.8972368Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8973066Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8973151Z graph_break [] 2025-12-04T09:50:15.8973332Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8973953Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8974057Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8974229Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8974316Z unimplemented [] 2025-12-04T09:50:15.8974458Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8975101Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8975188Z graph_break [] 2025-12-04T09:50:15.8975438Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8976061Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8976167Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8976344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8976432Z unimplemented [] 2025-12-04T09:50:15.8976568Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8977210Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8977347Z graph_break [] 2025-12-04T09:50:15.8977521Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8978144Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8978250Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8978462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8978550Z unimplemented [] 2025-12-04T09:50:15.8978690Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8979328Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8979417Z graph_break [] 2025-12-04T09:50:15.8979629Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8980257Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8980360Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8980536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8980622Z unimplemented [] 2025-12-04T09:50:15.8980762Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8981408Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8981493Z graph_break [] 2025-12-04T09:50:15.8981665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8982291Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8982397Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8982575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8982670Z unimplemented [] 2025-12-04T09:50:15.8982802Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8983452Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8983543Z graph_break [] 2025-12-04T09:50:15.8983714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8984330Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8984442Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8984618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8984751Z unimplemented [] 2025-12-04T09:50:15.8984886Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8985527Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8985619Z graph_break [] 2025-12-04T09:50:15.8985788Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8986411Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8986506Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8986678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8986810Z unimplemented [] 2025-12-04T09:50:15.8986941Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8987588Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8987675Z graph_break [] 2025-12-04T09:50:15.8987888Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8988516Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8988613Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8988783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8988916Z unimplemented [] 2025-12-04T09:50:15.8989047Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8989690Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8989781Z graph_break [] 2025-12-04T09:50:15.8989951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8990581Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8990681Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8990856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8990951Z unimplemented [] 2025-12-04T09:50:15.8991082Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8991734Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8991826Z graph_break [] 2025-12-04T09:50:15.8991995Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8992649Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8992769Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8992939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8993032Z unimplemented [] 2025-12-04T09:50:15.8993165Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8993811Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8993898Z graph_break [] 2025-12-04T09:50:15.8994070Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8994739Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8994836Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8995010Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8995107Z unimplemented [] 2025-12-04T09:50:15.8995238Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8995890Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8995978Z graph_break [] 2025-12-04T09:50:15.8996149Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8996821Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8996921Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8997096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8997225Z unimplemented [] 2025-12-04T09:50:15.8997359Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.8998014Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.8998101Z graph_break [] 2025-12-04T09:50:15.8998276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.8998940Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.8999044Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.8999230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.8999317Z unimplemented [] 2025-12-04T09:50:15.8999448Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9000104Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9000186Z graph_break [] 2025-12-04T09:50:15.9000360Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9000986Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9001086Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9001269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9001354Z unimplemented [] 2025-12-04T09:50:15.9001484Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9002129Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9002214Z graph_break [] 2025-12-04T09:50:15.9002391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9003025Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9003134Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9003339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9003427Z unimplemented [] 2025-12-04T09:50:15.9003602Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9004253Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9004338Z graph_break [] 2025-12-04T09:50:15.9004516Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9005136Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9005233Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9005412Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9005503Z unimplemented [] 2025-12-04T09:50:15.9005682Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9006331Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9006415Z graph_break [] 2025-12-04T09:50:15.9006595Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9007325Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9007425Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9007607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9007693Z unimplemented [] 2025-12-04T09:50:15.9007835Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9008553Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9008642Z graph_break [] 2025-12-04T09:50:15.9008822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9009445Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9009544Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9009716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9009804Z unimplemented [] 2025-12-04T09:50:15.9009939Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9010583Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9010670Z graph_break [] 2025-12-04T09:50:15.9010854Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9011471Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9011576Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9011748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9011833Z unimplemented [] 2025-12-04T09:50:15.9011969Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9012640Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9012744Z graph_break [] 2025-12-04T09:50:15.9012924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9013586Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9013690Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9013864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9013954Z unimplemented [] 2025-12-04T09:50:15.9014090Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9014732Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9014823Z graph_break [] 2025-12-04T09:50:15.9014994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9015615Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9015770Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9015942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9016029Z unimplemented [] 2025-12-04T09:50:15.9016209Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9016852Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9016944Z graph_break [] 2025-12-04T09:50:15.9017118Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9017743Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9017888Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9018108Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.9018215Z Traceback (most recent call last): 2025-12-04T09:50:15.9018597Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.9018680Z ).run(code) 2025-12-04T09:50:15.9018977Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9019064Z Searched string: 2025-12-04T09:50:15.9019148Z int32_t u0_raw; 2025-12-04T09:50:15.9019345Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9019429Z auto u0 = u0_raw; 2025-12-04T09:50:15.9019527Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9019799Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9020055Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9020339Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9020344Z 2025-12-04T09:50:15.9020475Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9020603Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9020717Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9021252Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9021392Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9021520Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9021645Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9021771Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9022253Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9022418Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9022689Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9022840Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9023239Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9023347Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9023452Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9023563Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9023567Z 2025-12-04T09:50:15.9023576Z 2025-12-04T09:50:15.9023580Z 2025-12-04T09:50:15.9023584Z 2025-12-04T09:50:15.9024307Z Wrapper code written to: /tmp/tmpdmenoyyn/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cemrlaz5akifwm2uwo55hfewjik7ptccekystgnh7oesdcu2bdpk.wrapper.cpp 2025-12-04T09:50:15.9024977Z Kernel code written to: /tmp/tmpdmenoyyn/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cfrjo6drc5g7e3fnmuuz4dhb6k2szgxox4zc67jksesenvs76g4p.kernel.cpp 2025-12-04T09:50:15.9025138Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9025185Z 2025-12-04T09:50:15.9025190Z 2025-12-04T09:50:15.9025380Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9025897Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9025901Z 2025-12-04T09:50:15.9026130Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9026354Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9026446Z unimplemented [] 2025-12-04T09:50:15.9026588Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9027159Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9027247Z graph_break [] 2025-12-04T09:50:15.9027428Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9028056Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9028155Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9028340Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9028428Z unimplemented [] 2025-12-04T09:50:15.9028570Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9029226Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9029309Z graph_break [] 2025-12-04T09:50:15.9029488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9030112Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9030214Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9030385Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9030472Z unimplemented [] 2025-12-04T09:50:15.9030611Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9031256Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9031342Z graph_break [] 2025-12-04T09:50:15.9031574Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9032195Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9032311Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9032486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9032572Z unimplemented [] 2025-12-04T09:50:15.9032707Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9033342Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9033467Z graph_break [] 2025-12-04T09:50:15.9033646Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9034270Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9034371Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9034584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9034674Z unimplemented [] 2025-12-04T09:50:15.9034813Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9036302Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9036393Z graph_break [] 2025-12-04T09:50:15.9036652Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9037278Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9037381Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9037554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9037644Z unimplemented [] 2025-12-04T09:50:15.9037788Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9038429Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9038518Z graph_break [] 2025-12-04T09:50:15.9038688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9039308Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9039421Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9039596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9039683Z unimplemented [] 2025-12-04T09:50:15.9039822Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9040466Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9040555Z graph_break [] 2025-12-04T09:50:15.9040730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9041350Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9041459Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9041630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9041790Z unimplemented [] 2025-12-04T09:50:15.9041924Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9042594Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9042695Z graph_break [] 2025-12-04T09:50:15.9042880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9043502Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9043609Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9043785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9043937Z unimplemented [] 2025-12-04T09:50:15.9044069Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9044713Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9044805Z graph_break [] 2025-12-04T09:50:15.9045066Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9045693Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9045790Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9045963Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9046097Z unimplemented [] 2025-12-04T09:50:15.9046229Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9046875Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9046966Z graph_break [] 2025-12-04T09:50:15.9047138Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9047828Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9047926Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9048102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9048200Z unimplemented [] 2025-12-04T09:50:15.9048332Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9048982Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9049080Z graph_break [] 2025-12-04T09:50:15.9049260Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9049885Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9049987Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9050157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9050243Z unimplemented [] 2025-12-04T09:50:15.9050380Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9051017Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9051113Z graph_break [] 2025-12-04T09:50:15.9051284Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9051950Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9052052Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9052225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9052331Z unimplemented [] 2025-12-04T09:50:15.9052479Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9053137Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9053224Z graph_break [] 2025-12-04T09:50:15.9053403Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9054071Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9054166Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9054338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9054432Z unimplemented [] 2025-12-04T09:50:15.9054606Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9055252Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9055343Z graph_break [] 2025-12-04T09:50:15.9055516Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9056183Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9056284Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9056456Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9056551Z unimplemented [] 2025-12-04T09:50:15.9056683Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9057336Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9057419Z graph_break [] 2025-12-04T09:50:15.9057591Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9058217Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9058318Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9058489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9058582Z unimplemented [] 2025-12-04T09:50:15.9058713Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9059361Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9059447Z graph_break [] 2025-12-04T09:50:15.9059617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9060242Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9060342Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9060523Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9060612Z unimplemented [] 2025-12-04T09:50:15.9060743Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9061426Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9061510Z graph_break [] 2025-12-04T09:50:15.9061683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9062315Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9062412Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9062592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9062680Z unimplemented [] 2025-12-04T09:50:15.9062855Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9063545Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9063632Z graph_break [] 2025-12-04T09:50:15.9063802Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9064479Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9064577Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9064755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9064839Z unimplemented [] 2025-12-04T09:50:15.9064968Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9065653Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9065741Z graph_break [] 2025-12-04T09:50:15.9065917Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9066541Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9066637Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9066815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9066900Z unimplemented [] 2025-12-04T09:50:15.9067030Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9067677Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9067764Z graph_break [] 2025-12-04T09:50:15.9067939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9068558Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9068652Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9068831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9068919Z unimplemented [] 2025-12-04T09:50:15.9069050Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9069691Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9069775Z graph_break [] 2025-12-04T09:50:15.9069951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9070616Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9070718Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9070893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9070980Z unimplemented [] 2025-12-04T09:50:15.9071119Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9071758Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9071840Z graph_break [] 2025-12-04T09:50:15.9072015Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9072641Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9072789Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9072960Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9073065Z unimplemented [] 2025-12-04T09:50:15.9073216Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9073902Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9073988Z graph_break [] 2025-12-04T09:50:15.9074169Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9074785Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9074927Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9075102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9075188Z unimplemented [] 2025-12-04T09:50:15.9075323Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9075968Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9076047Z graph_break [] 2025-12-04T09:50:15.9076223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9076840Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9076946Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9077118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9077204Z unimplemented [] 2025-12-04T09:50:15.9077341Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9077983Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9078072Z graph_break [] 2025-12-04T09:50:15.9078246Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9078864Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9078965Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9079136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9079224Z unimplemented [] 2025-12-04T09:50:15.9079370Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9080054Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9080146Z graph_break [] 2025-12-04T09:50:15.9080317Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9080940Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9081043Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9081213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9081298Z unimplemented [] 2025-12-04T09:50:15.9081436Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9082087Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9082241Z graph_break [] 2025-12-04T09:50:15.9082433Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9083119Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9083221Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9083393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9083483Z unimplemented [] 2025-12-04T09:50:15.9083616Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9084255Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9084391Z graph_break [] 2025-12-04T09:50:15.9084563Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9085181Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9085288Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9085458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9085554Z unimplemented [] 2025-12-04T09:50:15.9085684Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9086321Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9086416Z graph_break [] 2025-12-04T09:50:15.9086591Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9087218Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9087382Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9087554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9087649Z unimplemented [] 2025-12-04T09:50:15.9087781Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9088424Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9088512Z graph_break [] 2025-12-04T09:50:15.9088681Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9089357Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9089453Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9089623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9089713Z unimplemented [] 2025-12-04T09:50:15.9089847Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9090484Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9090571Z graph_break [] 2025-12-04T09:50:15.9090741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9091369Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9091509Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9091680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9091771Z unimplemented [] 2025-12-04T09:50:15.9091899Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9092583Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9092666Z graph_break [] 2025-12-04T09:50:15.9092850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9093518Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9093651Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9093823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9093921Z unimplemented [] 2025-12-04T09:50:15.9094054Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9094706Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9094789Z graph_break [] 2025-12-04T09:50:15.9094959Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9095590Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9095685Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9095861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9095950Z unimplemented [] 2025-12-04T09:50:15.9096079Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9096730Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9096816Z graph_break [] 2025-12-04T09:50:15.9096987Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9097616Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9097711Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9097885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9097970Z unimplemented [] 2025-12-04T09:50:15.9098102Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9098792Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9098873Z graph_break [] 2025-12-04T09:50:15.9099041Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9099671Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9099766Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9099942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9100030Z unimplemented [] 2025-12-04T09:50:15.9100160Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9100804Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9100928Z graph_break [] 2025-12-04T09:50:15.9101108Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9101726Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9101859Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9102037Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9102122Z unimplemented [] 2025-12-04T09:50:15.9102252Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9102912Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9103052Z graph_break [] 2025-12-04T09:50:15.9103252Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9103874Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9103970Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9104152Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9104238Z unimplemented [] 2025-12-04T09:50:15.9104368Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9105021Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9105106Z graph_break [] 2025-12-04T09:50:15.9105288Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9105914Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9106008Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9106188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9106273Z unimplemented [] 2025-12-04T09:50:15.9106412Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9107053Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9107136Z graph_break [] 2025-12-04T09:50:15.9107311Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9107928Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9108028Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9108251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9108338Z unimplemented [] 2025-12-04T09:50:15.9108479Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9109119Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9109199Z graph_break [] 2025-12-04T09:50:15.9109375Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9109995Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9110144Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9110352Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.9110464Z Traceback (most recent call last): 2025-12-04T09:50:15.9110846Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.9110924Z ).run(code) 2025-12-04T09:50:15.9111253Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9111345Z Searched string: 2025-12-04T09:50:15.9111424Z int32_t u0_raw; 2025-12-04T09:50:15.9111616Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9111711Z auto u0 = u0_raw; 2025-12-04T09:50:15.9111806Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9112076Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9112368Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9112704Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9112715Z 2025-12-04T09:50:15.9112841Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9112966Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9113084Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9113620Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9113746Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9113874Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9113998Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9114122Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9114560Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9114722Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9114998Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9115145Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9115535Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9115646Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9115749Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9115855Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9115859Z 2025-12-04T09:50:15.9115863Z 2025-12-04T09:50:15.9115867Z 2025-12-04T09:50:15.9115871Z 2025-12-04T09:50:15.9116550Z Wrapper code written to: /tmp/tmpcy3bphor/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/clue5griazmfemg7eie6jkmj47njrrty5gnqmbxotqaqrnayuoud.wrapper.cpp 2025-12-04T09:50:15.9117292Z Kernel code written to: /tmp/tmpcy3bphor/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cahydebz2qcbrplvppces6bue3ubuhcg67naceowbpa5uge3vv3s.kernel.cpp 2025-12-04T09:50:15.9117456Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9117461Z 2025-12-04T09:50:15.9117464Z 2025-12-04T09:50:15.9117647Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9118165Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9118170Z 2025-12-04T09:50:15.9118388Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9118567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9118658Z unimplemented [] 2025-12-04T09:50:15.9118830Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9119396Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9119480Z graph_break [] 2025-12-04T09:50:15.9119653Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9120322Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9120420Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9120599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9120687Z unimplemented [] 2025-12-04T09:50:15.9120817Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9121502Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9121592Z graph_break [] 2025-12-04T09:50:15.9121769Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9122391Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9122489Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9122666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9122752Z unimplemented [] 2025-12-04T09:50:15.9122882Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9123522Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9123609Z graph_break [] 2025-12-04T09:50:15.9123787Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9124407Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9124503Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9124684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9124771Z unimplemented [] 2025-12-04T09:50:15.9124901Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9125542Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9125627Z graph_break [] 2025-12-04T09:50:15.9125810Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9126475Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9126572Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9126751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9126839Z unimplemented [] 2025-12-04T09:50:15.9126979Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9127686Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9127772Z graph_break [] 2025-12-04T09:50:15.9127950Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9128577Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9128717Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9128895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9128980Z unimplemented [] 2025-12-04T09:50:15.9129114Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9129794Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9129882Z graph_break [] 2025-12-04T09:50:15.9130058Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9130679Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9130820Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9130995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9131081Z unimplemented [] 2025-12-04T09:50:15.9131217Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9131854Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9131937Z graph_break [] 2025-12-04T09:50:15.9132110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9132729Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9132835Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9133020Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9133121Z unimplemented [] 2025-12-04T09:50:15.9133287Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9133930Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9134011Z graph_break [] 2025-12-04T09:50:15.9134192Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9134813Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9134916Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9135085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9135172Z unimplemented [] 2025-12-04T09:50:15.9135577Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9136291Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9136385Z graph_break [] 2025-12-04T09:50:15.9136556Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9137175Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9137276Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9137445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9137530Z unimplemented [] 2025-12-04T09:50:15.9137664Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9138307Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9138459Z graph_break [] 2025-12-04T09:50:15.9138631Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9139304Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9139407Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9139576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9139671Z unimplemented [] 2025-12-04T09:50:15.9139802Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9140443Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9140592Z graph_break [] 2025-12-04T09:50:15.9140765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9141384Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9141486Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9141655Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9141748Z unimplemented [] 2025-12-04T09:50:15.9141880Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9142523Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9142616Z graph_break [] 2025-12-04T09:50:15.9142787Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9143406Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9143506Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9143677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9143770Z unimplemented [] 2025-12-04T09:50:15.9143904Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9144540Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9144626Z graph_break [] 2025-12-04T09:50:15.9144799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9145426Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9145568Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9145744Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9145838Z unimplemented [] 2025-12-04T09:50:15.9145974Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9146612Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9146701Z graph_break [] 2025-12-04T09:50:15.9146871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9147501Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9147640Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9147814Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9147904Z unimplemented [] 2025-12-04T09:50:15.9148034Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9148725Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9153613Z graph_break [] 2025-12-04T09:50:15.9153822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9154459Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9154659Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9154846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9154937Z unimplemented [] 2025-12-04T09:50:15.9155081Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9155732Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9155819Z graph_break [] 2025-12-04T09:50:15.9155994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9156624Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9156727Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9156907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9157001Z unimplemented [] 2025-12-04T09:50:15.9157136Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9157787Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9157873Z graph_break [] 2025-12-04T09:50:15.9158055Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9158673Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9158771Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9158952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9159038Z unimplemented [] 2025-12-04T09:50:15.9159174Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9159867Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9159954Z graph_break [] 2025-12-04T09:50:15.9160134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9160760Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9160858Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9161038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9161124Z unimplemented [] 2025-12-04T09:50:15.9161257Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9161909Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9162038Z graph_break [] 2025-12-04T09:50:15.9162225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9162897Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9163034Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9163216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9163303Z unimplemented [] 2025-12-04T09:50:15.9163444Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9164089Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9164214Z graph_break [] 2025-12-04T09:50:15.9164400Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9165021Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9165126Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9165300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9165388Z unimplemented [] 2025-12-04T09:50:15.9165527Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9166172Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9166254Z graph_break [] 2025-12-04T09:50:15.9166436Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9167062Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9167167Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9167418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9167507Z unimplemented [] 2025-12-04T09:50:15.9167647Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9168292Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9168373Z graph_break [] 2025-12-04T09:50:15.9168551Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9169169Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9169279Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9169499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9169587Z unimplemented [] 2025-12-04T09:50:15.9169726Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9170366Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9170456Z graph_break [] 2025-12-04T09:50:15.9170629Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9171251Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9171357Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9171568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9171655Z unimplemented [] 2025-12-04T09:50:15.9171795Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9172475Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9172569Z graph_break [] 2025-12-04T09:50:15.9172741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9173360Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9173466Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9173676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9173767Z unimplemented [] 2025-12-04T09:50:15.9173904Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9174547Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9174640Z graph_break [] 2025-12-04T09:50:15.9174816Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9175436Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9175539Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9175710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9175804Z unimplemented [] 2025-12-04T09:50:15.9175938Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9176587Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9176679Z graph_break [] 2025-12-04T09:50:15.9176849Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9177473Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9177577Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9177747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9177839Z unimplemented [] 2025-12-04T09:50:15.9177971Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9178615Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9178710Z graph_break [] 2025-12-04T09:50:15.9178926Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9179557Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9179654Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9179824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9179915Z unimplemented [] 2025-12-04T09:50:15.9180044Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9180679Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9180809Z graph_break [] 2025-12-04T09:50:15.9180982Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9181610Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9181704Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9181914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9182011Z unimplemented [] 2025-12-04T09:50:15.9182142Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9182826Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9182920Z graph_break [] 2025-12-04T09:50:15.9183132Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9183761Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9183860Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9184034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9184125Z unimplemented [] 2025-12-04T09:50:15.9184259Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9184904Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9184987Z graph_break [] 2025-12-04T09:50:15.9185157Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9185788Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9185888Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9186061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9186155Z unimplemented [] 2025-12-04T09:50:15.9186287Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9186938Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9187023Z graph_break [] 2025-12-04T09:50:15.9187195Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9187822Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9187927Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9188110Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9188241Z unimplemented [] 2025-12-04T09:50:15.9188374Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9189024Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9189108Z graph_break [] 2025-12-04T09:50:15.9189279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9189906Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9190005Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9190185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9190312Z unimplemented [] 2025-12-04T09:50:15.9190444Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9191099Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9191186Z graph_break [] 2025-12-04T09:50:15.9191422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9192052Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9192148Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9192334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9192460Z unimplemented [] 2025-12-04T09:50:15.9192591Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9193248Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9193331Z graph_break [] 2025-12-04T09:50:15.9193509Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9194133Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9194231Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9194411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9194497Z unimplemented [] 2025-12-04T09:50:15.9194633Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9195282Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9195371Z graph_break [] 2025-12-04T09:50:15.9195549Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9196171Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9196268Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9196449Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9196535Z unimplemented [] 2025-12-04T09:50:15.9196668Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9197313Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9197404Z graph_break [] 2025-12-04T09:50:15.9197581Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9198246Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9198342Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9198526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9198615Z unimplemented [] 2025-12-04T09:50:15.9198753Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9199398Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9199483Z graph_break [] 2025-12-04T09:50:15.9199662Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9200328Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9200425Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9200605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9200731Z unimplemented [] 2025-12-04T09:50:15.9200875Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9201513Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9201596Z graph_break [] 2025-12-04T09:50:15.9201773Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9202434Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9202543Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9202714Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9202801Z unimplemented [] 2025-12-04T09:50:15.9202943Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9203582Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9203663Z graph_break [] 2025-12-04T09:50:15.9203838Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9204454Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9204565Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9204738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9204829Z unimplemented [] 2025-12-04T09:50:15.9204972Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9205612Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9205696Z graph_break [] 2025-12-04T09:50:15.9205873Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9206493Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9206599Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9206770Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9206860Z unimplemented [] 2025-12-04T09:50:15.9207048Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9207769Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9207860Z graph_break [] 2025-12-04T09:50:15.9208039Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9208658Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9208760Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9208971Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.9209081Z Traceback (most recent call last): 2025-12-04T09:50:15.9209509Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 152, in test_unbacked_symint 2025-12-04T09:50:15.9209702Z ).check("RAIIAtenTensorHandle(tmp_tensor_handle_1);").run(code) 2025-12-04T09:50:15.9210359Z RuntimeError: Expected to find "AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_4, int_array_5, &tmp_tensor_handle_1));" but did not find it 2025-12-04T09:50:15.9210451Z Searched string: 2025-12-04T09:50:15.9210568Z AtenTensorHandle tmp_tensor_handle_0; 2025-12-04T09:50:15.9211007Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool1, 0, cached_torch_dtype_int32, 0, int_array_1, int_array_1, &tmp_tensor_handle_0)); 2025-12-04T09:50:15.9211170Z auto buf0 = RAIIAtenTensorHandle(tmp_tensor_handle_0); 2025-12-04T09:50:15.9211392Z // Topologically Sorted Source Nodes: [add], Original ATen: [aten.add] 2025-12-04T09:50:15.9211717Z call_triton_poi_fused_add_0(arg0_1, buf0, 1L, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9211804Z arg0_1.reset(); 2025-12-04T09:50:15.9211895Z int32_t u0_raw; 2025-12-04T09:50:15.9212083Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9212168Z auto u0 = u0_raw; 2025-12-04T09:50:15.9212275Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9212542Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9212815Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9213131Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9213137Z 2025-12-04T09:50:15.9213268Z const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9213406Z const int64_t int_array_3[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9213516Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9214070Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_2, int_array_3, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9214199Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9214314Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9214762Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_2, int_array_3, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9214922Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9215193Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9215354Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9215741Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9215864Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9215970Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9216115Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9216120Z 2025-12-04T09:50:15.9216124Z 2025-12-04T09:50:15.9216128Z 2025-12-04T09:50:15.9216132Z 2025-12-04T09:50:15.9216820Z Wrapper code written to: /tmp/tmpe9ow1ce4/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cp4wratawjoyn6veo7c7pniqcdshy3h7wnaeeq4q2ncvekzuu77f.wrapper.cpp 2025-12-04T09:50:15.9217474Z Kernel code written to: /tmp/tmpe9ow1ce4/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cwsdnf4b3yr5jok3w3xyd7td32qsmgywldxqo5oxz2ucpm2xkfqy.kernel.cpp 2025-12-04T09:50:15.9217958Z From CHECK: AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_4, int_array_5, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9217962Z 2025-12-04T09:50:15.9217968Z 2025-12-04T09:50:15.9218151Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9218720Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9218724Z 2025-12-04T09:50:15.9218948Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9219129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9219271Z unimplemented [] 2025-12-04T09:50:15.9219410Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9219982Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9220068Z graph_break [] 2025-12-04T09:50:15.9220243Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9220926Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9221026Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9221203Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9221296Z unimplemented [] 2025-12-04T09:50:15.9221428Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9222082Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9222167Z graph_break [] 2025-12-04T09:50:15.9222342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9223025Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9223130Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9223312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9223401Z unimplemented [] 2025-12-04T09:50:15.9223532Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9224190Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9224274Z graph_break [] 2025-12-04T09:50:15.9224447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9225081Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9225181Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9225365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9225453Z unimplemented [] 2025-12-04T09:50:15.9225589Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9226289Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9226374Z graph_break [] 2025-12-04T09:50:15.9226550Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9227182Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9227282Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9227461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9227551Z unimplemented [] 2025-12-04T09:50:15.9227683Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9228404Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9228486Z graph_break [] 2025-12-04T09:50:15.9228664Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9229328Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9229425Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9229605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9229690Z unimplemented [] 2025-12-04T09:50:15.9229820Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9230512Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9230602Z graph_break [] 2025-12-04T09:50:15.9230782Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9231406Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9231504Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9231683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9231770Z unimplemented [] 2025-12-04T09:50:15.9231904Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9232562Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9232653Z graph_break [] 2025-12-04T09:50:15.9232835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9233463Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9233564Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9233747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9233837Z unimplemented [] 2025-12-04T09:50:15.9233983Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9234629Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9234717Z graph_break [] 2025-12-04T09:50:15.9234901Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9235877Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9235981Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9236167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9236257Z unimplemented [] 2025-12-04T09:50:15.9236400Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9237044Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9237131Z graph_break [] 2025-12-04T09:50:15.9237315Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9237946Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9238120Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9238294Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9238381Z unimplemented [] 2025-12-04T09:50:15.9238524Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9239238Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9239328Z graph_break [] 2025-12-04T09:50:15.9239512Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9240137Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9240306Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9240486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9240573Z unimplemented [] 2025-12-04T09:50:15.9240716Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9241374Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9241456Z graph_break [] 2025-12-04T09:50:15.9241640Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9242264Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9242371Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9242547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9242640Z unimplemented [] 2025-12-04T09:50:15.9242789Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9243443Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9243536Z graph_break [] 2025-12-04T09:50:15.9243713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9244345Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9244450Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9244629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9244718Z unimplemented [] 2025-12-04T09:50:15.9244865Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9245565Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9245649Z graph_break [] 2025-12-04T09:50:15.9245831Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9246458Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9246554Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9246736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9246821Z unimplemented [] 2025-12-04T09:50:15.9246961Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9247662Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9247791Z graph_break [] 2025-12-04T09:50:15.9247969Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9248627Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9248733Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9248906Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9248998Z unimplemented [] 2025-12-04T09:50:15.9249137Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9249781Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9249915Z graph_break [] 2025-12-04T09:50:15.9250088Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9250713Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9250818Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9250992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9251081Z unimplemented [] 2025-12-04T09:50:15.9251222Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9251869Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9251963Z graph_break [] 2025-12-04T09:50:15.9252144Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9252826Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9252933Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9253112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9253211Z unimplemented [] 2025-12-04T09:50:15.9253346Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9253993Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9254087Z graph_break [] 2025-12-04T09:50:15.9254263Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9254888Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9255036Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9255210Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9255302Z unimplemented [] 2025-12-04T09:50:15.9255434Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9256079Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9256168Z graph_break [] 2025-12-04T09:50:15.9256339Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9256967Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9257106Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9257281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9257378Z unimplemented [] 2025-12-04T09:50:15.9257511Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9258188Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9258278Z graph_break [] 2025-12-04T09:50:15.9258447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9259080Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9259215Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9259389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9259484Z unimplemented [] 2025-12-04T09:50:15.9259618Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9260259Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9260352Z graph_break [] 2025-12-04T09:50:15.9260522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9261150Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9261247Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9261417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9261517Z unimplemented [] 2025-12-04T09:50:15.9261647Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9262303Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9262402Z graph_break [] 2025-12-04T09:50:15.9262598Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9263229Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9263328Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9263499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9263589Z unimplemented [] 2025-12-04T09:50:15.9263724Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9264451Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9264534Z graph_break [] 2025-12-04T09:50:15.9264707Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9265346Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9265445Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9265630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9265716Z unimplemented [] 2025-12-04T09:50:15.9265849Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9266502Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9266629Z graph_break [] 2025-12-04T09:50:15.9266805Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9267436Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9267578Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9267760Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9267849Z unimplemented [] 2025-12-04T09:50:15.9267979Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9268628Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9268747Z graph_break [] 2025-12-04T09:50:15.9268922Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9269553Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9269653Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9269831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9269917Z unimplemented [] 2025-12-04T09:50:15.9270047Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9270695Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9270777Z graph_break [] 2025-12-04T09:50:15.9270964Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9271594Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9271689Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9271873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9271958Z unimplemented [] 2025-12-04T09:50:15.9272090Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9272796Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9272881Z graph_break [] 2025-12-04T09:50:15.9273062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9273684Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9273788Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9274011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9274101Z unimplemented [] 2025-12-04T09:50:15.9274232Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9274884Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9274968Z graph_break [] 2025-12-04T09:50:15.9275151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9275780Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9275882Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9276103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9276195Z unimplemented [] 2025-12-04T09:50:15.9276336Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9277018Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9277102Z graph_break [] 2025-12-04T09:50:15.9277285Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9277905Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9278006Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9278227Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9278320Z unimplemented [] 2025-12-04T09:50:15.9278457Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9279105Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9279188Z graph_break [] 2025-12-04T09:50:15.9279379Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9280005Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9280114Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9280285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9280373Z unimplemented [] 2025-12-04T09:50:15.9280511Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9281162Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9281247Z graph_break [] 2025-12-04T09:50:15.9281426Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9282058Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9282160Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9282338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9282427Z unimplemented [] 2025-12-04T09:50:15.9282572Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9283220Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9283305Z graph_break [] 2025-12-04T09:50:15.9283526Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9284156Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9284263Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9284438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9284527Z unimplemented [] 2025-12-04T09:50:15.9284671Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9285316Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9285452Z graph_break [] 2025-12-04T09:50:15.9285623Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9286248Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9286348Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9286555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9286641Z unimplemented [] 2025-12-04T09:50:15.9286778Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9287477Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9287564Z graph_break [] 2025-12-04T09:50:15.9287775Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9288408Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9288511Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9288685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9288771Z unimplemented [] 2025-12-04T09:50:15.9288911Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9289551Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9289641Z graph_break [] 2025-12-04T09:50:15.9289812Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9290434Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9290535Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9290710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9290803Z unimplemented [] 2025-12-04T09:50:15.9290933Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9291585Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9291677Z graph_break [] 2025-12-04T09:50:15.9291846Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9292467Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9292574Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9292745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9292903Z unimplemented [] 2025-12-04T09:50:15.9293054Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9293710Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9293798Z graph_break [] 2025-12-04T09:50:15.9293971Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9294602Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9294701Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9294875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9295008Z unimplemented [] 2025-12-04T09:50:15.9295144Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9295796Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9295884Z graph_break [] 2025-12-04T09:50:15.9296091Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9296727Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9296823Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9297001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9297160Z unimplemented [] 2025-12-04T09:50:15.9297292Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9297941Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9298034Z graph_break [] 2025-12-04T09:50:15.9298207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9298838Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9298936Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9299114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9299209Z unimplemented [] 2025-12-04T09:50:15.9299342Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9299995Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9300085Z graph_break [] 2025-12-04T09:50:15.9300262Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9300896Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9300995Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9301172Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9301272Z unimplemented [] 2025-12-04T09:50:15.9301402Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9302065Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9302150Z graph_break [] 2025-12-04T09:50:15.9302366Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9302999Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9303095Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9303277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9303362Z unimplemented [] 2025-12-04T09:50:15.9303496Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9304155Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9304239Z graph_break [] 2025-12-04T09:50:15.9304411Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9305090Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9305186Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9305366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9305491Z unimplemented [] 2025-12-04T09:50:15.9305625Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9306271Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9306356Z graph_break [] 2025-12-04T09:50:15.9306528Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9307195Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9307296Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9307516Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.9307621Z Traceback (most recent call last): 2025-12-04T09:50:15.9307998Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.9308085Z ).run(code) 2025-12-04T09:50:15.9308372Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9308463Z Searched string: 2025-12-04T09:50:15.9308545Z int32_t u0_raw; 2025-12-04T09:50:15.9308737Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9308833Z auto u0 = u0_raw; 2025-12-04T09:50:15.9308932Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9309195Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9309462Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9309741Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9309747Z 2025-12-04T09:50:15.9309880Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9310011Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9310119Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9310672Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9310798Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9310931Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9311057Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9311172Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9311661Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9311826Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9312097Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9312253Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9312637Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9312751Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9312855Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9312959Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9313002Z 2025-12-04T09:50:15.9313006Z 2025-12-04T09:50:15.9313010Z 2025-12-04T09:50:15.9313014Z 2025-12-04T09:50:15.9313744Z Wrapper code written to: /tmp/tmpdn8ehesu/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c3434vpuducveivzgy45l3hnwmbbkmk5qzhh4f5dh2ch2ql5yn3h.wrapper.cpp 2025-12-04T09:50:15.9314406Z Kernel code written to: /tmp/tmpdn8ehesu/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cxevdyjgulfdlre3fog5bqdygt5za3neqcp4kc4qlng6jgmrggjl.kernel.cpp 2025-12-04T09:50:15.9314612Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9314617Z 2025-12-04T09:50:15.9314621Z 2025-12-04T09:50:15.9314802Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9315321Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9315335Z 2025-12-04T09:50:15.9315591Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9315773Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9315868Z unimplemented [] 2025-12-04T09:50:15.9316004Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9316575Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9316670Z graph_break [] 2025-12-04T09:50:15.9316846Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9317481Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9317582Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9317755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9317849Z unimplemented [] 2025-12-04T09:50:15.9317984Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9318635Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9318722Z graph_break [] 2025-12-04T09:50:15.9318897Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9319529Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9319631Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9319809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9319903Z unimplemented [] 2025-12-04T09:50:15.9320038Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9320736Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9320822Z graph_break [] 2025-12-04T09:50:15.9320994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9321624Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9321724Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9321899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9321990Z unimplemented [] 2025-12-04T09:50:15.9322123Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9322776Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9322904Z graph_break [] 2025-12-04T09:50:15.9323083Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9323765Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9323900Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9324079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9324167Z unimplemented [] 2025-12-04T09:50:15.9324300Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9324946Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9325069Z graph_break [] 2025-12-04T09:50:15.9325245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9325877Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9325973Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9326154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9326242Z unimplemented [] 2025-12-04T09:50:15.9326371Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9327027Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9327111Z graph_break [] 2025-12-04T09:50:15.9327331Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9327970Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9328070Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9328253Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9328340Z unimplemented [] 2025-12-04T09:50:15.9328474Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9329126Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9329207Z graph_break [] 2025-12-04T09:50:15.9329388Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9330011Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9330114Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9330343Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9330431Z unimplemented [] 2025-12-04T09:50:15.9330564Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9331217Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9331301Z graph_break [] 2025-12-04T09:50:15.9331482Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9332106Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9332207Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9332426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9332511Z unimplemented [] 2025-12-04T09:50:15.9332643Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9333360Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9333446Z graph_break [] 2025-12-04T09:50:15.9333625Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9334244Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9334339Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9334559Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9334647Z unimplemented [] 2025-12-04T09:50:15.9334782Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9335645Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9335730Z graph_break [] 2025-12-04T09:50:15.9335908Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9336532Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9336632Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9336805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9336890Z unimplemented [] 2025-12-04T09:50:15.9337028Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9337678Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9337759Z graph_break [] 2025-12-04T09:50:15.9337938Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9338567Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9338670Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9338841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9338925Z unimplemented [] 2025-12-04T09:50:15.9339059Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9339703Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9339787Z graph_break [] 2025-12-04T09:50:15.9340050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9340678Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9340779Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9340950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9341034Z unimplemented [] 2025-12-04T09:50:15.9341171Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9341816Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9341968Z graph_break [] 2025-12-04T09:50:15.9342139Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9342765Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9342866Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9343093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9343182Z unimplemented [] 2025-12-04T09:50:15.9343321Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9343970Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9344059Z graph_break [] 2025-12-04T09:50:15.9344286Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9344917Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9345017Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9345188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9345277Z unimplemented [] 2025-12-04T09:50:15.9345427Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9346072Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9350491Z graph_break [] 2025-12-04T09:50:15.9350687Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9351332Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9351437Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9351617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9351713Z unimplemented [] 2025-12-04T09:50:15.9351851Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9352500Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9352595Z graph_break [] 2025-12-04T09:50:15.9352770Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9353396Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9353496Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9353673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9353830Z unimplemented [] 2025-12-04T09:50:15.9353965Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9354614Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9354700Z graph_break [] 2025-12-04T09:50:15.9354873Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9355500Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9355599Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9355775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9355907Z unimplemented [] 2025-12-04T09:50:15.9356042Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9356690Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9356773Z graph_break [] 2025-12-04T09:50:15.9356982Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9357606Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9357704Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9357875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9358008Z unimplemented [] 2025-12-04T09:50:15.9358139Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9358792Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9358875Z graph_break [] 2025-12-04T09:50:15.9359046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9359674Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9359771Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9359947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9360033Z unimplemented [] 2025-12-04T09:50:15.9360165Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9360816Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9360903Z graph_break [] 2025-12-04T09:50:15.9361076Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9361714Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9361813Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9361992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9362078Z unimplemented [] 2025-12-04T09:50:15.9362211Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9362858Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9362947Z graph_break [] 2025-12-04T09:50:15.9363125Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9363790Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9363888Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9364073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9364159Z unimplemented [] 2025-12-04T09:50:15.9364291Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9364941Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9365026Z graph_break [] 2025-12-04T09:50:15.9365201Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9365864Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9365961Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9366139Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9366263Z unimplemented [] 2025-12-04T09:50:15.9366394Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9367046Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9367133Z graph_break [] 2025-12-04T09:50:15.9367401Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9368062Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9368168Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9368345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9368433Z unimplemented [] 2025-12-04T09:50:15.9368571Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9369214Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9369297Z graph_break [] 2025-12-04T09:50:15.9369476Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9370094Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9370196Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9370373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9370463Z unimplemented [] 2025-12-04T09:50:15.9370599Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9371240Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9371322Z graph_break [] 2025-12-04T09:50:15.9371498Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9372115Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9372222Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9372393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9372480Z unimplemented [] 2025-12-04T09:50:15.9372664Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9373308Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9373390Z graph_break [] 2025-12-04T09:50:15.9373573Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9374190Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9374291Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9374460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9374550Z unimplemented [] 2025-12-04T09:50:15.9374755Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9375402Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9375483Z graph_break [] 2025-12-04T09:50:15.9375660Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9376326Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9376429Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9376599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9376685Z unimplemented [] 2025-12-04T09:50:15.9376821Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9377517Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9377612Z graph_break [] 2025-12-04T09:50:15.9377784Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9378404Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9378506Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9378676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9378762Z unimplemented [] 2025-12-04T09:50:15.9378898Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9379537Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9379630Z graph_break [] 2025-12-04T09:50:15.9379803Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9380421Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9380525Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9380696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9380789Z unimplemented [] 2025-12-04T09:50:15.9380921Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9381565Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9381659Z graph_break [] 2025-12-04T09:50:15.9381830Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9382504Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9382626Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9382822Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9382917Z unimplemented [] 2025-12-04T09:50:15.9383048Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9383687Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9383779Z graph_break [] 2025-12-04T09:50:15.9383953Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9384576Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9384724Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9384896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9384989Z unimplemented [] 2025-12-04T09:50:15.9385121Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9385797Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9385890Z graph_break [] 2025-12-04T09:50:15.9386065Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9386691Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9386832Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9387004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9387099Z unimplemented [] 2025-12-04T09:50:15.9387230Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9387876Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9387967Z graph_break [] 2025-12-04T09:50:15.9388136Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9388763Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9388865Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9389041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9389134Z unimplemented [] 2025-12-04T09:50:15.9389265Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9389908Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9389998Z graph_break [] 2025-12-04T09:50:15.9390168Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9390795Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9390891Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9391062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9391155Z unimplemented [] 2025-12-04T09:50:15.9391290Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9391981Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9392065Z graph_break [] 2025-12-04T09:50:15.9392241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9392864Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9392960Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9393131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9393225Z unimplemented [] 2025-12-04T09:50:15.9393361Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9394055Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9394137Z graph_break [] 2025-12-04T09:50:15.9394308Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9394981Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9395080Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9395263Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9395353Z unimplemented [] 2025-12-04T09:50:15.9395484Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9396137Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9396265Z graph_break [] 2025-12-04T09:50:15.9396443Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9397075Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9397175Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9397356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9397445Z unimplemented [] 2025-12-04T09:50:15.9397579Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9398239Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9398327Z graph_break [] 2025-12-04T09:50:15.9398501Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9399138Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9399234Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9399421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9399511Z unimplemented [] 2025-12-04T09:50:15.9399640Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9400287Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9400372Z graph_break [] 2025-12-04T09:50:15.9400552Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9401219Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9401315Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9401498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9401585Z unimplemented [] 2025-12-04T09:50:15.9401716Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9402368Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9402471Z graph_break [] 2025-12-04T09:50:15.9402675Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9403290Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9403428Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9403612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9403702Z unimplemented [] 2025-12-04T09:50:15.9403835Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9404529Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9404611Z graph_break [] 2025-12-04T09:50:15.9404795Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9405417Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9405552Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9405729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9405818Z unimplemented [] 2025-12-04T09:50:15.9405966Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9406611Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9406693Z graph_break [] 2025-12-04T09:50:15.9406878Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9407552Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9407659Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9407833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9407923Z unimplemented [] 2025-12-04T09:50:15.9408063Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9408711Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9408796Z graph_break [] 2025-12-04T09:50:15.9408983Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9409606Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9409710Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9409923Z ___________________ TestMemoryPlanning.test_unbacked_symint ____________________ 2025-12-04T09:50:15.9410034Z Traceback (most recent call last): 2025-12-04T09:50:15.9410420Z File "/var/lib/jenkins/workspace/test/inductor/test_memory_planning.py", line 140, in test_unbacked_symint 2025-12-04T09:50:15.9410506Z ).run(code) 2025-12-04T09:50:15.9410869Z RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9410961Z Searched string: 2025-12-04T09:50:15.9411046Z int32_t u0_raw; 2025-12-04T09:50:15.9411247Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9411335Z auto u0 = u0_raw; 2025-12-04T09:50:15.9411433Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9411704Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9411954Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9412229Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9412244Z 2025-12-04T09:50:15.9412435Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9412579Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9412705Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9413246Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9413420Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9413555Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9413686Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9413814Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9414256Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9414453Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9414737Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9414891Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9415283Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9415402Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9415510Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9415618Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9415623Z 2025-12-04T09:50:15.9415627Z 2025-12-04T09:50:15.9415631Z 2025-12-04T09:50:15.9415635Z 2025-12-04T09:50:15.9416317Z Wrapper code written to: /tmp/tmpftd7gr4m/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/chvvm7pqxonb6jnagw6kaleehrv4ipxgtaqtu7cv3msevmhhyyuk.wrapper.cpp 2025-12-04T09:50:15.9416989Z Kernel code written to: /tmp/tmpftd7gr4m/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cmew4ywcgkrg6feynwpvs2uav3eqrkmhq2kmvonbi5xaossp5pcj.kernel.cpp 2025-12-04T09:50:15.9417160Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9417164Z 2025-12-04T09:50:15.9417168Z 2025-12-04T09:50:15.9417350Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9417874Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9417881Z 2025-12-04T09:50:15.9418101Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9418294Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9418386Z unimplemented [] 2025-12-04T09:50:15.9418522Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9419097Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9419187Z graph_break [] 2025-12-04T09:50:15.9419365Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9420047Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9420149Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9420336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9420422Z unimplemented [] 2025-12-04T09:50:15.9420558Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9421217Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9421299Z graph_break [] 2025-12-04T09:50:15.9421486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9422174Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9422276Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9422494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9422599Z unimplemented [] 2025-12-04T09:50:15.9422774Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9423426Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9423511Z graph_break [] 2025-12-04T09:50:15.9423687Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9424344Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9424448Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9424630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9424719Z unimplemented [] 2025-12-04T09:50:15.9424848Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9425505Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9425584Z graph_break [] 2025-12-04T09:50:15.9425766Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9426389Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9426493Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9426677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9426770Z unimplemented [] 2025-12-04T09:50:15.9426907Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9427555Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9427636Z graph_break [] 2025-12-04T09:50:15.9427814Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9428436Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9428544Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9428719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9428809Z unimplemented [] 2025-12-04T09:50:15.9428956Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9429648Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9429730Z graph_break [] 2025-12-04T09:50:15.9429913Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9430539Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9430647Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9430827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9430921Z unimplemented [] 2025-12-04T09:50:15.9431068Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9431758Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9431841Z graph_break [] 2025-12-04T09:50:15.9432025Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9432682Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9432794Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9432971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9433059Z unimplemented [] 2025-12-04T09:50:15.9433202Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9433882Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9433978Z graph_break [] 2025-12-04T09:50:15.9434153Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9434779Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9434890Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9435068Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9435156Z unimplemented [] 2025-12-04T09:50:15.9435591Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9436242Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9436345Z graph_break [] 2025-12-04T09:50:15.9436522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9437151Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9437264Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9437446Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9437530Z unimplemented [] 2025-12-04T09:50:15.9437674Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9438322Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9438413Z graph_break [] 2025-12-04T09:50:15.9438590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9439296Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9439407Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9439582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9439686Z unimplemented [] 2025-12-04T09:50:15.9439817Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9440464Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9440563Z graph_break [] 2025-12-04T09:50:15.9440739Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9441368Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9441538Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9441713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9441800Z unimplemented [] 2025-12-04T09:50:15.9441935Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9442670Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9442766Z graph_break [] 2025-12-04T09:50:15.9442949Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9443574Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9443770Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9443949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9444038Z unimplemented [] 2025-12-04T09:50:15.9444183Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9444825Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9444920Z graph_break [] 2025-12-04T09:50:15.9445093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9445711Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9445822Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9445993Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9446087Z unimplemented [] 2025-12-04T09:50:15.9446219Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9446858Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9446956Z graph_break [] 2025-12-04T09:50:15.9447131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9447835Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9447942Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9448117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9448211Z unimplemented [] 2025-12-04T09:50:15.9448345Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9449030Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9449124Z graph_break [] 2025-12-04T09:50:15.9449295Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9449927Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9450027Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9450201Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9450292Z unimplemented [] 2025-12-04T09:50:15.9450422Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9451060Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9451195Z graph_break [] 2025-12-04T09:50:15.9451369Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9452034Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9452133Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9452306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9452404Z unimplemented [] 2025-12-04T09:50:15.9452537Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9453186Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9453316Z graph_break [] 2025-12-04T09:50:15.9453491Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9454117Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9454217Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9454392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9454490Z unimplemented [] 2025-12-04T09:50:15.9454622Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9455271Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9455360Z graph_break [] 2025-12-04T09:50:15.9455533Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9456163Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9456260Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9456431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9456523Z unimplemented [] 2025-12-04T09:50:15.9456653Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9457308Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9457393Z graph_break [] 2025-12-04T09:50:15.9457566Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9458197Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9458340Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9458525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9458613Z unimplemented [] 2025-12-04T09:50:15.9458749Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9459404Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9459487Z graph_break [] 2025-12-04T09:50:15.9459658Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9460284Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9460424Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9460611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9460695Z unimplemented [] 2025-12-04T09:50:15.9460825Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9461517Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9461604Z graph_break [] 2025-12-04T09:50:15.9461780Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9462408Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9462541Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9462723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9462815Z unimplemented [] 2025-12-04T09:50:15.9462951Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9463597Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9463681Z graph_break [] 2025-12-04T09:50:15.9463858Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9464478Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9464574Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9464762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9464855Z unimplemented [] 2025-12-04T09:50:15.9464991Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9465642Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9465728Z graph_break [] 2025-12-04T09:50:15.9465919Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9466542Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9466640Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9466816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9466903Z unimplemented [] 2025-12-04T09:50:15.9467037Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9467735Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9467816Z graph_break [] 2025-12-04T09:50:15.9467995Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9468621Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9468718Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9468898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9468982Z unimplemented [] 2025-12-04T09:50:15.9469121Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9469767Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9469894Z graph_break [] 2025-12-04T09:50:15.9470077Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9470695Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9470835Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9471011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9471100Z unimplemented [] 2025-12-04T09:50:15.9471238Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9471877Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9471998Z graph_break [] 2025-12-04T09:50:15.9472189Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9472809Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9472913Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9473095Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9473180Z unimplemented [] 2025-12-04T09:50:15.9473318Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9474015Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9474095Z graph_break [] 2025-12-04T09:50:15.9474276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9474901Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9475010Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9475183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9475268Z unimplemented [] 2025-12-04T09:50:15.9475410Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9476051Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9476136Z graph_break [] 2025-12-04T09:50:15.9476313Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9476936Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9477045Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9477262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9477351Z unimplemented [] 2025-12-04T09:50:15.9477498Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9478143Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9478236Z graph_break [] 2025-12-04T09:50:15.9478404Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9479025Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9479131Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9479344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9479434Z unimplemented [] 2025-12-04T09:50:15.9479573Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9480272Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9480362Z graph_break [] 2025-12-04T09:50:15.9480535Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9481156Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9481263Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9481472Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9481569Z unimplemented [] 2025-12-04T09:50:15.9481705Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9482359Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9482450Z graph_break [] 2025-12-04T09:50:15.9482624Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9483282Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9483403Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9483578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9483679Z unimplemented [] 2025-12-04T09:50:15.9483810Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9484456Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9484547Z graph_break [] 2025-12-04T09:50:15.9484721Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9485350Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9485451Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9485624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9485713Z unimplemented [] 2025-12-04T09:50:15.9485845Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9486488Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9486581Z graph_break [] 2025-12-04T09:50:15.9486803Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9487489Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9487585Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9487755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9487850Z unimplemented [] 2025-12-04T09:50:15.9487980Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9488621Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9488753Z graph_break [] 2025-12-04T09:50:15.9488923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9489556Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9489655Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9489870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9489960Z unimplemented [] 2025-12-04T09:50:15.9490090Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9490734Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9490816Z graph_break [] 2025-12-04T09:50:15.9491031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9491666Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9491761Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9491931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9492025Z unimplemented [] 2025-12-04T09:50:15.9492158Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9492812Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9492895Z graph_break [] 2025-12-04T09:50:15.9493067Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9493700Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9493800Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9493986Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9494072Z unimplemented [] 2025-12-04T09:50:15.9494209Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9494866Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9494949Z graph_break [] 2025-12-04T09:50:15.9495124Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9495755Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9495857Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9496040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9496178Z unimplemented [] 2025-12-04T09:50:15.9496314Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9496974Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9497061Z graph_break [] 2025-12-04T09:50:15.9497234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9497867Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9497968Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9498166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9498297Z unimplemented [] 2025-12-04T09:50:15.9498434Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9499104Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9499189Z graph_break [] 2025-12-04T09:50:15.9499418Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9500042Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9500140Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9500326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9500451Z unimplemented [] 2025-12-04T09:50:15.9500582Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9501255Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9501340Z graph_break [] 2025-12-04T09:50:15.9501519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9502142Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9502239Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9502422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9502512Z unimplemented [] 2025-12-04T09:50:15.9502651Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9503306Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9503394Z graph_break [] 2025-12-04T09:50:15.9503603Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9504245Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9504344Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9504535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9504622Z unimplemented [] 2025-12-04T09:50:15.9504770Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9505410Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9505500Z graph_break [] 2025-12-04T09:50:15.9505739Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9506368Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9506467Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9506650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9506734Z unimplemented [] 2025-12-04T09:50:15.9506875Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9507522Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9507610Z graph_break [] 2025-12-04T09:50:15.9507800Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9508463Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9508575Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9508757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9508884Z unimplemented [] 2025-12-04T09:50:15.9509034Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9509677Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9509764Z graph_break [] 2025-12-04T09:50:15.9509945Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9510615Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9510729Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9510901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9510986Z unimplemented [] 2025-12-04T09:50:15.9511123Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9511767Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9511849Z graph_break [] 2025-12-04T09:50:15.9512027Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9512651Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9512763Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9512941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:50:15.9513030Z unimplemented [] 2025-12-04T09:50:15.9513176Z stats [('calls_captured', 9), ('unique_graphs', 1)] 2025-12-04T09:50:15.9513863Z inductor [('extern_calls', 8), ('async_compile_cache_miss', 6), ('async_compile_cache_hit', 2), ('benchmarking.InductorBenchmarker.benchmark', 2), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2)] 2025-12-04T09:50:15.9513959Z graph_break [] 2025-12-04T09:50:15.9514134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:50:15.9514755Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:50:15.9514861Z return cls.__new__(cls, *args) 2025-12-04T09:50:15.9515476Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_memory_planning/inductor.test_memory_planning-7b0b7b5cd4d35939.xml - 2025-12-04T09:50:15.9515681Z =========================== short test summary info ============================ 2025-12-04T09:50:15.9516326Z FAILED [5.2639s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9516418Z Searched string: 2025-12-04T09:50:15.9516512Z int32_t u0_raw; 2025-12-04T09:50:15.9516702Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9516790Z auto u0 = u0_raw; 2025-12-04T09:50:15.9516898Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9517165Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9517432Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9517788Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9517794Z 2025-12-04T09:50:15.9517926Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9518064Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9518178Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9518766Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9518894Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9519024Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9519166Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9519285Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9519769Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9519943Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9520228Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9520390Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9520792Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9520899Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9521012Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9521116Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9521120Z 2025-12-04T09:50:15.9521124Z 2025-12-04T09:50:15.9521128Z 2025-12-04T09:50:15.9521132Z 2025-12-04T09:50:15.9521822Z Wrapper code written to: /tmp/tmpynmakc3o/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cvnxn3bizurhuzr7gkhjhxjjk4hyhhvhrzwo6jq6lrtygjt46rir.wrapper.cpp 2025-12-04T09:50:15.9522500Z Kernel code written to: /tmp/tmpynmakc3o/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cpixalai6zyzrpagcxp6pl7wnxudmwznsxpme4ci47jq24gmyg45.kernel.cpp 2025-12-04T09:50:15.9522665Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9522670Z 2025-12-04T09:50:15.9522680Z 2025-12-04T09:50:15.9522865Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9523382Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9523387Z 2025-12-04T09:50:15.9523612Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9524261Z FAILED [5.3353s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9524361Z Searched string: 2025-12-04T09:50:15.9524446Z int32_t u0_raw; 2025-12-04T09:50:15.9524679Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9524778Z auto u0 = u0_raw; 2025-12-04T09:50:15.9524877Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9525140Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9525401Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9525686Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9525691Z 2025-12-04T09:50:15.9525828Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9525959Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9526071Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9526660Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9526791Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9526929Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9527057Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9527298Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9527754Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9527916Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9528199Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9528356Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9528788Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9528912Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9529020Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9529130Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9529135Z 2025-12-04T09:50:15.9529139Z 2025-12-04T09:50:15.9529142Z 2025-12-04T09:50:15.9529146Z 2025-12-04T09:50:15.9529829Z Wrapper code written to: /tmp/tmpw7pv0svu/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cgjwlds4j4fhy2xfxkgjbedsxcycy76i57x4bu7vps3eezkb7ayc.wrapper.cpp 2025-12-04T09:50:15.9530493Z Kernel code written to: /tmp/tmpw7pv0svu/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cbmy3vm5t27asmqclrs5sldgfynirjxyqx2gevp5suefbzlhnf23.kernel.cpp 2025-12-04T09:50:15.9530668Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9530672Z 2025-12-04T09:50:15.9530679Z 2025-12-04T09:50:15.9530862Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9531392Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9531396Z 2025-12-04T09:50:15.9531623Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9532263Z FAILED [4.9371s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9532364Z Searched string: 2025-12-04T09:50:15.9532447Z int32_t u0_raw; 2025-12-04T09:50:15.9532661Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9532774Z auto u0 = u0_raw; 2025-12-04T09:50:15.9532882Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9533152Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9533415Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9533738Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9533751Z 2025-12-04T09:50:15.9533879Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9534007Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9534124Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9534673Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9534802Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9534932Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9535055Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9535176Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9535896Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9536054Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9536332Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9536548Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9536936Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9537050Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9537154Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9537259Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9537264Z 2025-12-04T09:50:15.9537267Z 2025-12-04T09:50:15.9537328Z 2025-12-04T09:50:15.9537332Z 2025-12-04T09:50:15.9538009Z Wrapper code written to: /tmp/tmp3dda8dk0/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cupi7kwsy66cim4u25rel2co425mz3wnowd2vpxdktnqsbsu5zqf.wrapper.cpp 2025-12-04T09:50:15.9538680Z Kernel code written to: /tmp/tmp3dda8dk0/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/czs3f4ouyqoo7qfzf7xfiovltvjemtk4oyoudnzs7f5bbnk6nrsl.kernel.cpp 2025-12-04T09:50:15.9538840Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9538844Z 2025-12-04T09:50:15.9538850Z 2025-12-04T09:50:15.9539032Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9539554Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9539559Z 2025-12-04T09:50:15.9539781Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9540422Z FAILED [4.9457s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9540507Z Searched string: 2025-12-04T09:50:15.9540589Z int32_t u0_raw; 2025-12-04T09:50:15.9540784Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9540865Z auto u0 = u0_raw; 2025-12-04T09:50:15.9540962Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9541229Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9541478Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9541764Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9541769Z 2025-12-04T09:50:15.9541894Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9546497Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9546625Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9547286Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9547428Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9547559Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9547699Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9547819Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9548266Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9548434Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9548706Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9548863Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9549315Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9549426Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9549540Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9549644Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9549649Z 2025-12-04T09:50:15.9549653Z 2025-12-04T09:50:15.9549698Z 2025-12-04T09:50:15.9549702Z 2025-12-04T09:50:15.9550388Z Wrapper code written to: /tmp/tmpxe3wnqs2/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/crpj7dz4k2hf7bipjm5e4dg4cbxzmks7bedsgusoswkzhijp2lci.wrapper.cpp 2025-12-04T09:50:15.9551063Z Kernel code written to: /tmp/tmpxe3wnqs2/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cc6edmxjttd2x3npllhthyultqraw3zsxaqevsnpgo3xoqne75wy.kernel.cpp 2025-12-04T09:50:15.9551228Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9551270Z 2025-12-04T09:50:15.9551277Z 2025-12-04T09:50:15.9551468Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9551997Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9552002Z 2025-12-04T09:50:15.9552232Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9553206Z FAILED [4.9427s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_4, int_array_5, &tmp_tensor_handle_1));" but did not find it 2025-12-04T09:50:15.9553293Z Searched string: 2025-12-04T09:50:15.9553413Z AtenTensorHandle tmp_tensor_handle_0; 2025-12-04T09:50:15.9553849Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool1, 0, cached_torch_dtype_int32, 0, int_array_1, int_array_1, &tmp_tensor_handle_0)); 2025-12-04T09:50:15.9554021Z auto buf0 = RAIIAtenTensorHandle(tmp_tensor_handle_0); 2025-12-04T09:50:15.9554238Z // Topologically Sorted Source Nodes: [add], Original ATen: [aten.add] 2025-12-04T09:50:15.9554529Z call_triton_poi_fused_add_0(arg0_1, buf0, 1L, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9554617Z arg0_1.reset(); 2025-12-04T09:50:15.9554701Z int32_t u0_raw; 2025-12-04T09:50:15.9554888Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9554987Z auto u0 = u0_raw; 2025-12-04T09:50:15.9555089Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9555356Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9555614Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9555899Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9555905Z 2025-12-04T09:50:15.9556044Z const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9556248Z const int64_t int_array_3[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9556363Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9556908Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_2, int_array_3, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9557035Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9557152Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9557597Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_2, int_array_3, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9557756Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9558039Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9558234Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9558642Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9558753Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9558858Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9559011Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9559016Z 2025-12-04T09:50:15.9559019Z 2025-12-04T09:50:15.9559023Z 2025-12-04T09:50:15.9559027Z 2025-12-04T09:50:15.9559696Z Wrapper code written to: /tmp/tmp8jct58wp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/chl47nn5sum3xy4ll77za6csl3dyhq2qf7k4l34y22amiydicuos.wrapper.cpp 2025-12-04T09:50:15.9560366Z Kernel code written to: /tmp/tmp8jct58wp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/czvwj7hef5hiutsegg3svscv3l3zfgefsx5senos3t2jm7jj2idu.kernel.cpp 2025-12-04T09:50:15.9560881Z From CHECK: AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_4, int_array_5, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9560887Z 2025-12-04T09:50:15.9560894Z 2025-12-04T09:50:15.9561086Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9561608Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9561612Z 2025-12-04T09:50:15.9561837Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9562483Z FAILED [4.9631s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9562567Z Searched string: 2025-12-04T09:50:15.9562664Z int32_t u0_raw; 2025-12-04T09:50:15.9562883Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9562978Z auto u0 = u0_raw; 2025-12-04T09:50:15.9563079Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9563338Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9563588Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9563875Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9563880Z 2025-12-04T09:50:15.9564007Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9564137Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9564242Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9564786Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9564923Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9565051Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9565228Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9565343Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9565783Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9565950Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9566224Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9566382Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9566765Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9566870Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9567019Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9567122Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9567126Z 2025-12-04T09:50:15.9567130Z 2025-12-04T09:50:15.9567136Z 2025-12-04T09:50:15.9567140Z 2025-12-04T09:50:15.9567883Z Wrapper code written to: /tmp/tmpus4s7msp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c32rinomxhjvwr4rzfv5wrndcnn2avny2ug5lapit3q3lecrwi2u.wrapper.cpp 2025-12-04T09:50:15.9568584Z Kernel code written to: /tmp/tmpus4s7msp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cxisfpmlgnp3b7qua4c5xvnmkaejdeklhn23wbpjytiwkgdzqp4i.kernel.cpp 2025-12-04T09:50:15.9568746Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9568751Z 2025-12-04T09:50:15.9568754Z 2025-12-04T09:50:15.9568947Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9569464Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9569509Z 2025-12-04T09:50:15.9569746Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9570386Z FAILED [4.9668s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9570472Z Searched string: 2025-12-04T09:50:15.9570562Z int32_t u0_raw; 2025-12-04T09:50:15.9570745Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9570832Z auto u0 = u0_raw; 2025-12-04T09:50:15.9570929Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9571196Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9571454Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9571733Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9571740Z 2025-12-04T09:50:15.9571872Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9572002Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9572109Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9572689Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9572838Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9572962Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9573093Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9573210Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9573654Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9573821Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9574137Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9574292Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9574676Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9574788Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9574891Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9574995Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9574999Z 2025-12-04T09:50:15.9575003Z 2025-12-04T09:50:15.9575006Z 2025-12-04T09:50:15.9575010Z 2025-12-04T09:50:15.9575694Z Wrapper code written to: /tmp/tmpo1v55153/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cuvl6siq5dkkznzwmgn2eken45qtj6syxgc5feulqbhinj5yo5y3.wrapper.cpp 2025-12-04T09:50:15.9576351Z Kernel code written to: /tmp/tmpo1v55153/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cgalohs3wmwvvu37brmwwzxo2ep75c73kqjgmlnxadvlio3itq24.kernel.cpp 2025-12-04T09:50:15.9576561Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9576566Z 2025-12-04T09:50:15.9576569Z 2025-12-04T09:50:15.9576752Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9577307Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9577318Z 2025-12-04T09:50:15.9577541Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9578172Z FAILED [4.9431s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9578305Z Searched string: 2025-12-04T09:50:15.9578384Z int32_t u0_raw; 2025-12-04T09:50:15.9578573Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9578661Z auto u0 = u0_raw; 2025-12-04T09:50:15.9578758Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9579026Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9579279Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9579557Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9579562Z 2025-12-04T09:50:15.9579692Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9579818Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9579936Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9580475Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9580609Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9580740Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9580866Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9580981Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9581431Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9581591Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9581864Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9582014Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9582405Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9582517Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9582622Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9582770Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9582780Z 2025-12-04T09:50:15.9582784Z 2025-12-04T09:50:15.9582787Z 2025-12-04T09:50:15.9582791Z 2025-12-04T09:50:15.9583477Z Wrapper code written to: /tmp/tmpojoo8738/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/czglehddruugo4yh3rb2m74ahflahffybpnjoz4vnhomsq4fdi3l.wrapper.cpp 2025-12-04T09:50:15.9584131Z Kernel code written to: /tmp/tmpojoo8738/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cygljwezmp44v7se3tjmyf3sk3ugun22b2uuh2smv557py6opwk6.kernel.cpp 2025-12-04T09:50:15.9584299Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9584304Z 2025-12-04T09:50:15.9584307Z 2025-12-04T09:50:15.9584489Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9585015Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9585058Z 2025-12-04T09:50:15.9585284Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9585922Z FAILED [4.9541s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9586041Z Searched string: 2025-12-04T09:50:15.9586122Z int32_t u0_raw; 2025-12-04T09:50:15.9586315Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9586400Z auto u0 = u0_raw; 2025-12-04T09:50:15.9586498Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9586769Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9587083Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9587366Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9587374Z 2025-12-04T09:50:15.9587499Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9587623Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9587740Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9588280Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9588414Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9588543Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9588671Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9588793Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9589236Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9589396Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9589675Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9589825Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9590227Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9590334Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9590441Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9590548Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9590553Z 2025-12-04T09:50:15.9590557Z 2025-12-04T09:50:15.9590561Z 2025-12-04T09:50:15.9590564Z 2025-12-04T09:50:15.9591244Z Wrapper code written to: /tmp/tmpldcbptk5/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cb6bbvswsmkgcpryj77u7wue5dmvexmfbg52j4qaepk5ljzqq3jf.wrapper.cpp 2025-12-04T09:50:15.9591960Z Kernel code written to: /tmp/tmpldcbptk5/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c2jafjcq5hjdfypuqff7zhhjgnfo7h247dkgrv44qnjxi2bqxbq2.kernel.cpp 2025-12-04T09:50:15.9592127Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9592131Z 2025-12-04T09:50:15.9592135Z 2025-12-04T09:50:15.9592324Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9592838Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9592843Z 2025-12-04T09:50:15.9593070Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9593759Z FAILED [4.9506s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9593883Z Searched string: 2025-12-04T09:50:15.9593967Z int32_t u0_raw; 2025-12-04T09:50:15.9594153Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9594234Z auto u0 = u0_raw; 2025-12-04T09:50:15.9594341Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9594598Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9594886Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9595165Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9595170Z 2025-12-04T09:50:15.9595295Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9595424Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9595998Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9596541Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9596685Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9596814Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9596949Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9597065Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9597501Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9597666Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9597934Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9598081Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9598474Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9598581Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9598688Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9598792Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9598796Z 2025-12-04T09:50:15.9598800Z 2025-12-04T09:50:15.9598804Z 2025-12-04T09:50:15.9598807Z 2025-12-04T09:50:15.9599475Z Wrapper code written to: /tmp/tmpzy568zuz/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c6ahe4lfy2mnl5l36m23pjwmp7cwevf3c4hjgspxzeyevquk733e.wrapper.cpp 2025-12-04T09:50:15.9600130Z Kernel code written to: /tmp/tmpzy568zuz/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/caoclm3q33lmmn5c6i4rhj6gnh64s3lxnggraykdmhbttsyvdgff.kernel.cpp 2025-12-04T09:50:15.9600288Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9600292Z 2025-12-04T09:50:15.9600298Z 2025-12-04T09:50:15.9600479Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9601044Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9601049Z 2025-12-04T09:50:15.9601272Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9601909Z FAILED [4.9997s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9601992Z Searched string: 2025-12-04T09:50:15.9602086Z int32_t u0_raw; 2025-12-04T09:50:15.9602274Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9602356Z auto u0 = u0_raw; 2025-12-04T09:50:15.9602462Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9602759Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9603076Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9603364Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9603369Z 2025-12-04T09:50:15.9603496Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9603628Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9603781Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9604331Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9604457Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9604585Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9604717Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9604869Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9605316Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9605474Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9605745Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9605899Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9606283Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9606392Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9606497Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9606599Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9606603Z 2025-12-04T09:50:15.9606607Z 2025-12-04T09:50:15.9606615Z 2025-12-04T09:50:15.9606619Z 2025-12-04T09:50:15.9607354Z Wrapper code written to: /tmp/tmpw2qoh7sg/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cjpjdca5j24wpvshtns3kzrfb2ssk6aorqlrbhz32uxsyism3epr.wrapper.cpp 2025-12-04T09:50:15.9608030Z Kernel code written to: /tmp/tmpw2qoh7sg/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/csmsledbnqmvnyxfs3la7tloykqs6vbf5ijravu3atekvmx3vgsr.kernel.cpp 2025-12-04T09:50:15.9608196Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9608200Z 2025-12-04T09:50:15.9608206Z 2025-12-04T09:50:15.9608381Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9608895Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9608899Z 2025-12-04T09:50:15.9609126Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9609758Z FAILED [4.9588s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9609852Z Searched string: 2025-12-04T09:50:15.9609932Z int32_t u0_raw; 2025-12-04T09:50:15.9610159Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9610247Z auto u0 = u0_raw; 2025-12-04T09:50:15.9610345Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9610601Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9610858Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9611133Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9611138Z 2025-12-04T09:50:15.9611269Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9611395Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9611539Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9612089Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9612215Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9612341Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9612501Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9612613Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9613056Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9613212Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9613482Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9613666Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9614055Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9614167Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9614270Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9614370Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9614374Z 2025-12-04T09:50:15.9614378Z 2025-12-04T09:50:15.9614386Z 2025-12-04T09:50:15.9614391Z 2025-12-04T09:50:15.9615073Z Wrapper code written to: /tmp/tmpvxygh32y/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cgpl4og5ixv4rtug3nalbf5qwfe36jepzefotlyqiwruuiefqv2u.wrapper.cpp 2025-12-04T09:50:15.9615735Z Kernel code written to: /tmp/tmpvxygh32y/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cwopjd5pu2bkgij4beg677lnpvemwrpe5a7bmotya6popxocixzo.kernel.cpp 2025-12-04T09:50:15.9615897Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9615904Z 2025-12-04T09:50:15.9615910Z 2025-12-04T09:50:15.9616086Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9616610Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9616615Z 2025-12-04T09:50:15.9616833Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9617471Z FAILED [4.9615s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9617561Z Searched string: 2025-12-04T09:50:15.9617644Z int32_t u0_raw; 2025-12-04T09:50:15.9617828Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9617913Z auto u0 = u0_raw; 2025-12-04T09:50:15.9618009Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9618271Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9618563Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9618845Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9618850Z 2025-12-04T09:50:15.9618974Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9619098Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9619206Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9619752Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9619877Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9620007Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9620131Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9620288Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9620727Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9620882Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9621213Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9621361Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9621751Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9621854Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9621959Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9622066Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9622112Z 2025-12-04T09:50:15.9622116Z 2025-12-04T09:50:15.9622122Z 2025-12-04T09:50:15.9622126Z 2025-12-04T09:50:15.9622800Z Wrapper code written to: /tmp/tmp8wgeup53/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cssyumg2n6nv2zh7npzbe4eykua5afxnllekqk3334hoi6mt7drz.wrapper.cpp 2025-12-04T09:50:15.9623513Z Kernel code written to: /tmp/tmp8wgeup53/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cvpxczncsdjy7i7w3xvcpnxvjq64wmxmn7aqpxszdtt3hkmbtskb.kernel.cpp 2025-12-04T09:50:15.9623674Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9623678Z 2025-12-04T09:50:15.9623682Z 2025-12-04T09:50:15.9623859Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9624379Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9624383Z 2025-12-04T09:50:15.9624600Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9625242Z FAILED [4.9727s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9625329Z Searched string: 2025-12-04T09:50:15.9625413Z int32_t u0_raw; 2025-12-04T09:50:15.9625600Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9625679Z auto u0 = u0_raw; 2025-12-04T09:50:15.9625785Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9626045Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9626292Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9626576Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9626580Z 2025-12-04T09:50:15.9626711Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9626843Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9626947Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9627530Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9627662Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9627787Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9627913Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9628031Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9628466Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9628632Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9628900Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9629089Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9629479Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9629584Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9629684Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9629786Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9629829Z 2025-12-04T09:50:15.9629834Z 2025-12-04T09:50:15.9629837Z 2025-12-04T09:50:15.9629841Z 2025-12-04T09:50:15.9630516Z Wrapper code written to: /tmp/tmpelvvkc93/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/ci2y3y2ehbgmq6urvxudpfsweevy453uwvzjgceaj4o5ucqbk2z5.wrapper.cpp 2025-12-04T09:50:15.9631187Z Kernel code written to: /tmp/tmpelvvkc93/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/czm3pxdnswipucfhbg3lvcx3vckszqd6ejlvqyeaz24xekalsqdm.kernel.cpp 2025-12-04T09:50:15.9631388Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9631395Z 2025-12-04T09:50:15.9631399Z 2025-12-04T09:50:15.9631579Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9632100Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9632105Z 2025-12-04T09:50:15.9632332Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9633022Z FAILED [5.1814s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9633105Z Searched string: 2025-12-04T09:50:15.9633190Z int32_t u0_raw; 2025-12-04T09:50:15.9633379Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9633467Z auto u0 = u0_raw; 2025-12-04T09:50:15.9633567Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9633821Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9634076Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9634347Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9634351Z 2025-12-04T09:50:15.9634476Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9634615Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9634727Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9635514Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9635646Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9635774Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9635898Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9636016Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9636526Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9636688Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9636957Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9637101Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9637495Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9637598Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9637696Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9637804Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9637865Z 2025-12-04T09:50:15.9637869Z 2025-12-04T09:50:15.9637873Z 2025-12-04T09:50:15.9637877Z 2025-12-04T09:50:15.9638561Z Wrapper code written to: /tmp/tmpmnablfyn/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/ccdx57szd4hcekqnkyjtrsxg3fyrjyzdc6xhkrzd5uxxzt7cpgnz.wrapper.cpp 2025-12-04T09:50:15.9639285Z Kernel code written to: /tmp/tmpmnablfyn/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cfneaoqujuwu67slsls2zwmniprcvyggjanfxxsts547ol7itigh.kernel.cpp 2025-12-04T09:50:15.9639444Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9639448Z 2025-12-04T09:50:15.9639452Z 2025-12-04T09:50:15.9639631Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9640146Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9640204Z 2025-12-04T09:50:15.9640427Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9641063Z FAILED [5.0192s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9641146Z Searched string: 2025-12-04T09:50:15.9641231Z int32_t u0_raw; 2025-12-04T09:50:15.9641414Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9641496Z auto u0 = u0_raw; 2025-12-04T09:50:15.9641598Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9641853Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9642106Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9642376Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9642384Z 2025-12-04T09:50:15.9642519Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9642644Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9642754Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9643293Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9643418Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9643541Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9643671Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9643784Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9644221Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9644380Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9644655Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9644848Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9645231Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9645335Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9645446Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9645546Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9645550Z 2025-12-04T09:50:15.9645554Z 2025-12-04T09:50:15.9645558Z 2025-12-04T09:50:15.9645561Z 2025-12-04T09:50:15.9646248Z Wrapper code written to: /tmp/tmpkylini8i/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cwczcj7eicdl2qcsmcnmshfpeahnrahk6ra7vqxgqmlrtddgoesg.wrapper.cpp 2025-12-04T09:50:15.9646898Z Kernel code written to: /tmp/tmpkylini8i/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cmn4evgqb3ylk4lnpi6wpfikeou73h5xiyv3o7sehn7t2g272d6w.kernel.cpp 2025-12-04T09:50:15.9647102Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9647106Z 2025-12-04T09:50:15.9647110Z 2025-12-04T09:50:15.9647330Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9647845Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9647892Z 2025-12-04T09:50:15.9648118Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9648748Z FAILED [4.9976s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9648835Z Searched string: 2025-12-04T09:50:15.9648912Z int32_t u0_raw; 2025-12-04T09:50:15.9649134Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9649224Z auto u0 = u0_raw; 2025-12-04T09:50:15.9649321Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9649578Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9649830Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9650105Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9650110Z 2025-12-04T09:50:15.9650237Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9650360Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9650462Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9651003Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9651129Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9651258Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9651381Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9651496Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9651934Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9652090Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9652362Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9652507Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9652890Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9653002Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9653106Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9653207Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9653211Z 2025-12-04T09:50:15.9653215Z 2025-12-04T09:50:15.9653222Z 2025-12-04T09:50:15.9653271Z 2025-12-04T09:50:15.9653928Z Wrapper code written to: /tmp/tmpc0i_15m8/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cen6wyrz2ysaa3zsesy4as76nzvmcin7ihspqhylnz7rclhppx2z.wrapper.cpp 2025-12-04T09:50:15.9654565Z Kernel code written to: /tmp/tmpc0i_15m8/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cey7dmpyvk26nj2nubouhfg6rccew7qvuurlrdrtb5vozh5chmfa.kernel.cpp 2025-12-04T09:50:15.9654731Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9654736Z 2025-12-04T09:50:15.9654739Z 2025-12-04T09:50:15.9654915Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9655432Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9655504Z 2025-12-04T09:50:15.9655722Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9656357Z FAILED [4.9392s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9656443Z Searched string: 2025-12-04T09:50:15.9656521Z int32_t u0_raw; 2025-12-04T09:50:15.9656751Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9656832Z auto u0 = u0_raw; 2025-12-04T09:50:15.9656929Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9657186Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9657433Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9657746Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9657753Z 2025-12-04T09:50:15.9657876Z const int64_t int_array_4[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9658001Z const int64_t int_array_5[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9658110Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9658649Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_4, int_array_5, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9658773Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9658897Z const int64_t int_array_6[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9659019Z const int64_t int_array_7[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9659134Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9659571Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_6, int_array_7, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9659731Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9660003Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9660148Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9660531Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9660635Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9660734Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9660836Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9660840Z 2025-12-04T09:50:15.9660844Z 2025-12-04T09:50:15.9660847Z 2025-12-04T09:50:15.9660851Z 2025-12-04T09:50:15.9661529Z Wrapper code written to: /tmp/tmpn2z2gjfp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/clew3iphn6yh3tmx5hxgcs3rcauyxhzlzdsa4ftdm6pr7tqy4oav.wrapper.cpp 2025-12-04T09:50:15.9662198Z Kernel code written to: /tmp/tmpn2z2gjfp/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cynr3jtbwdo5ke3poekmiwdi7jlwki3jvricsnupij5ddug7jorz.kernel.cpp 2025-12-04T09:50:15.9662400Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9662404Z 2025-12-04T09:50:15.9662408Z 2025-12-04T09:50:15.9662585Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9663101Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9663106Z 2025-12-04T09:50:15.9663323Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9664008Z FAILED [4.9643s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9664088Z Searched string: 2025-12-04T09:50:15.9664170Z int32_t u0_raw; 2025-12-04T09:50:15.9664395Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9664476Z auto u0 = u0_raw; 2025-12-04T09:50:15.9664575Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9664831Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9665078Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9665394Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9665399Z 2025-12-04T09:50:15.9665528Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9665658Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9665762Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9666298Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9666467Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9666590Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9666712Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9666827Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9667265Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9667423Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9667692Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9667836Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9668220Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9668328Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9668428Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9668530Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9668536Z 2025-12-04T09:50:15.9668540Z 2025-12-04T09:50:15.9668543Z 2025-12-04T09:50:15.9668547Z 2025-12-04T09:50:15.9669213Z Wrapper code written to: /tmp/tmp2prvqs7d/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c4g2b4ohpfuu43y73weckxfatc2nb7y7lk74z6qou2k4bk7uomxo.wrapper.cpp 2025-12-04T09:50:15.9669883Z Kernel code written to: /tmp/tmp2prvqs7d/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cbgblnxqluncaedgthvheljmtzs5tfg2cy3awug4lgqjepjrp7am.kernel.cpp 2025-12-04T09:50:15.9670040Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9670045Z 2025-12-04T09:50:15.9670049Z 2025-12-04T09:50:15.9670230Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9670747Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9670755Z 2025-12-04T09:50:15.9671016Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9671648Z FAILED [4.9370s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9671737Z Searched string: 2025-12-04T09:50:15.9671820Z int32_t u0_raw; 2025-12-04T09:50:15.9672002Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9672081Z auto u0 = u0_raw; 2025-12-04T09:50:15.9672180Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9672434Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9672686Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9673004Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9673010Z 2025-12-04T09:50:15.9673161Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9673315Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9673418Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9674004Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9674128Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9674252Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9674380Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9674493Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9674968Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9675132Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9675407Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9675556Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9675940Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9676043Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9676146Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9676245Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9676249Z 2025-12-04T09:50:15.9676253Z 2025-12-04T09:50:15.9676257Z 2025-12-04T09:50:15.9676260Z 2025-12-04T09:50:15.9676943Z Wrapper code written to: /tmp/tmpdmenoyyn/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cemrlaz5akifwm2uwo55hfewjik7ptccekystgnh7oesdcu2bdpk.wrapper.cpp 2025-12-04T09:50:15.9677602Z Kernel code written to: /tmp/tmpdmenoyyn/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cfrjo6drc5g7e3fnmuuz4dhb6k2szgxox4zc67jksesenvs76g4p.kernel.cpp 2025-12-04T09:50:15.9677759Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9677768Z 2025-12-04T09:50:15.9677772Z 2025-12-04T09:50:15.9677948Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9678461Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9678466Z 2025-12-04T09:50:15.9678687Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9679317Z FAILED [4.9368s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9679406Z Searched string: 2025-12-04T09:50:15.9679486Z int32_t u0_raw; 2025-12-04T09:50:15.9679672Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9679799Z auto u0 = u0_raw; 2025-12-04T09:50:15.9679895Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9680147Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9680400Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9680675Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9680680Z 2025-12-04T09:50:15.9680810Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9680934Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9681037Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9681579Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9681746Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9681871Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9681993Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9682107Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9682584Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9682742Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9683007Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9683155Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9683575Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9683688Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9683790Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9683889Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9683894Z 2025-12-04T09:50:15.9683897Z 2025-12-04T09:50:15.9683901Z 2025-12-04T09:50:15.9683905Z 2025-12-04T09:50:15.9684593Z Wrapper code written to: /tmp/tmpcy3bphor/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/clue5griazmfemg7eie6jkmj47njrrty5gnqmbxotqaqrnayuoud.wrapper.cpp 2025-12-04T09:50:15.9685253Z Kernel code written to: /tmp/tmpcy3bphor/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cahydebz2qcbrplvppces6bue3ubuhcg67naceowbpa5uge3vv3s.kernel.cpp 2025-12-04T09:50:15.9685416Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9685420Z 2025-12-04T09:50:15.9685424Z 2025-12-04T09:50:15.9685601Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9686118Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9686125Z 2025-12-04T09:50:15.9686340Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9687346Z FAILED [4.9397s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_4, int_array_5, &tmp_tensor_handle_1));" but did not find it 2025-12-04T09:50:15.9687440Z Searched string: 2025-12-04T09:50:15.9687552Z AtenTensorHandle tmp_tensor_handle_0; 2025-12-04T09:50:15.9687982Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool1, 0, cached_torch_dtype_int32, 0, int_array_1, int_array_1, &tmp_tensor_handle_0)); 2025-12-04T09:50:15.9688144Z auto buf0 = RAIIAtenTensorHandle(tmp_tensor_handle_0); 2025-12-04T09:50:15.9688356Z // Topologically Sorted Source Nodes: [add], Original ATen: [aten.add] 2025-12-04T09:50:15.9688716Z call_triton_poi_fused_add_0(arg0_1, buf0, 1L, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9688797Z arg0_1.reset(); 2025-12-04T09:50:15.9688875Z int32_t u0_raw; 2025-12-04T09:50:15.9689062Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9689141Z auto u0 = u0_raw; 2025-12-04T09:50:15.9689244Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9689494Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9689742Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9690020Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9690027Z 2025-12-04T09:50:15.9690150Z const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9690323Z const int64_t int_array_3[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9690427Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9690965Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_2, int_array_3, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9691094Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9691244Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9691678Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_2, int_array_3, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9691841Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9692107Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9692294Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9692677Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9692783Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9692884Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9692984Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9692989Z 2025-12-04T09:50:15.9692993Z 2025-12-04T09:50:15.9692996Z 2025-12-04T09:50:15.9693000Z 2025-12-04T09:50:15.9693683Z Wrapper code written to: /tmp/tmpe9ow1ce4/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cp4wratawjoyn6veo7c7pniqcdshy3h7wnaeeq4q2ncvekzuu77f.wrapper.cpp 2025-12-04T09:50:15.9694335Z Kernel code written to: /tmp/tmpe9ow1ce4/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cwsdnf4b3yr5jok3w3xyd7td32qsmgywldxqo5oxz2ucpm2xkfqy.kernel.cpp 2025-12-04T09:50:15.9694815Z From CHECK: AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_4, int_array_5, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9694824Z 2025-12-04T09:50:15.9694827Z 2025-12-04T09:50:15.9695007Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9695520Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9695525Z 2025-12-04T09:50:15.9695747Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9696376Z FAILED [4.9458s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9696463Z Searched string: 2025-12-04T09:50:15.9696541Z int32_t u0_raw; 2025-12-04T09:50:15.9696724Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9696810Z auto u0 = u0_raw; 2025-12-04T09:50:15.9696909Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9697164Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9697460Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9697735Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9697739Z 2025-12-04T09:50:15.9697872Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9697993Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9698097Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9698639Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9698763Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9698889Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9699051Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9699163Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9699607Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9699762Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9700076Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9700222Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9700606Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9700714Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9700815Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9700955Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9700961Z 2025-12-04T09:50:15.9700965Z 2025-12-04T09:50:15.9700972Z 2025-12-04T09:50:15.9700976Z 2025-12-04T09:50:15.9701646Z Wrapper code written to: /tmp/tmpdn8ehesu/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/c3434vpuducveivzgy45l3hnwmbbkmk5qzhh4f5dh2ch2ql5yn3h.wrapper.cpp 2025-12-04T09:50:15.9702306Z Kernel code written to: /tmp/tmpdn8ehesu/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cxevdyjgulfdlre3fog5bqdygt5za3neqcp4kc4qlng6jgmrggjl.kernel.cpp 2025-12-04T09:50:15.9702471Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9702475Z 2025-12-04T09:50:15.9702479Z 2025-12-04T09:50:15.9702655Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9703168Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9703172Z 2025-12-04T09:50:15.9703393Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9704029Z FAILED [4.9531s] inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint - RuntimeError: Expected to find "const int64_t int_array_2[] = {10L, 8L*u0, 32L};" but did not find it 2025-12-04T09:50:15.9704117Z Searched string: 2025-12-04T09:50:15.9704195Z int32_t u0_raw; 2025-12-04T09:50:15.9704380Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_item_int32(buf0, &u0_raw)); 2025-12-04T09:50:15.9704460Z auto u0 = u0_raw; 2025-12-04T09:50:15.9704556Z pool1.reset(); buf0.reset(); 2025-12-04T09:50:15.9704813Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9705061Z if (!(1L <= u0)) { throw std::runtime_error("Expected 1 <= u0 but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9705337Z if (!(u0 >= 1L)) { throw std::runtime_error("Expected u0 >= 1 to be True but received " + std::to_string(u0)); } 2025-12-04T09:50:15.9705344Z 2025-12-04T09:50:15.9705469Z const int64_t int_array_3[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9705595Z const int64_t int_array_4[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9705745Z AtenTensorHandle pool0_handle; 2025-12-04T09:50:15.9706285Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch_empty_strided(3, int_array_3, int_array_4, cached_torch_dtype_float32, cached_torch_device_type_cuda, this->device_idx_, &pool0_handle)); 2025-12-04T09:50:15.9706410Z RAIIAtenTensorHandle pool0(pool0_handle); 2025-12-04T09:50:15.9706537Z const int64_t int_array_5[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9706658Z const int64_t int_array_6[] = {256L*u0, 32L, 1L}; 2025-12-04T09:50:15.9706782Z AtenTensorHandle tmp_tensor_handle_1; 2025-12-04T09:50:15.9707214Z AOTI_TORCH_ERROR_CODE_CHECK(aoti_torch__alloc_from_pool(pool0, 0, cached_torch_dtype_float32, 3, int_array_5, int_array_6, &tmp_tensor_handle_1)); 2025-12-04T09:50:15.9707369Z auto buf5 = RAIIAtenTensorHandle(tmp_tensor_handle_1); 2025-12-04T09:50:15.9707645Z // Topologically Sorted Source Nodes: [fill_, mul_2], Original ATen: [aten.fill, aten.mul] 2025-12-04T09:50:15.9707831Z int64_t triton_poi_fused_fill_mul_1_xnumel = 2560L*u0; 2025-12-04T09:50:15.9708221Z call_triton_poi_fused_fill_mul_1(buf5, triton_poi_fused_fill_mul_1_xnumel, this->device_idx_, stream, kernels, this->cubin_dir_); 2025-12-04T09:50:15.9708325Z output_handles[0] = buf5.release(); 2025-12-04T09:50:15.9708424Z } // AOTInductorModel::run_impl 2025-12-04T09:50:15.9708564Z } // namespace torch::aot_inductor 2025-12-04T09:50:15.9708569Z 2025-12-04T09:50:15.9708572Z 2025-12-04T09:50:15.9708576Z 2025-12-04T09:50:15.9708580Z 2025-12-04T09:50:15.9709259Z Wrapper code written to: /tmp/tmpftd7gr4m/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/chvvm7pqxonb6jnagw6kaleehrv4ipxgtaqtu7cv3msevmhhyyuk.wrapper.cpp 2025-12-04T09:50:15.9709927Z Kernel code written to: /tmp/tmpftd7gr4m/cgpzuvaqovcgu2cfnmq4lduy3y3v5jrpsv25z52bm7ghrgdpu4wq/cmew4ywcgkrg6feynwpvs2uav3eqrkmhq2kmvonbi5xaossp5pcj.kernel.cpp 2025-12-04T09:50:15.9710126Z From CHECK: const int64_t int_array_2[] = {10L, 8L*u0, 32L}; 2025-12-04T09:50:15.9710131Z 2025-12-04T09:50:15.9710134Z 2025-12-04T09:50:15.9710315Z To execute this test, run the following from the base repo dir: 2025-12-04T09:50:15.9710837Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/inductor/test_memory_planning.py TestMemoryPlanning.test_unbacked_symint 2025-12-04T09:50:15.9710842Z 2025-12-04T09:50:15.9711060Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:50:15.9711219Z ================== 24 failed, 26 passed in 252.20s (0:04:12) =================== 2025-12-04T09:50:15.9711224Z 2025-12-04T09:50:15.9711686Z FINISHED PRINTING LOG FILE of inductor/test_memory_planning 1/1 (test/test-reports/inductor.test_memory_planning_1.1_ab660c17d0398f84_.log) 2025-12-04T09:50:15.9711691Z 2025-12-04T09:50:15.9711984Z Finished inductor/test_memory_planning 1/1 ... [2025-12-04 09:50:15.768436][2318.184902193], took 4.33min 2025-12-04T09:50:15.9712646Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_memory_planning/inductor.test_memory_planning-7b0b7b5cd4d35939.xml 2025-12-04T09:50:16.5023742Z Uploading logs for 57118183167 to S3 2025-12-04T09:50:16.5533436Z Uploading artifacts took 0.70 seconds 2025-12-04T09:50:16.5533785Z inductor/test_memory_planning 1/1 failed! 2025-12-04T09:50:16.5538326Z Running export/test_cpp_serdes 1/1 ... [2025-12-04 09:50:16.553502][2318.96997329] 2025-12-04T09:50:16.5538892Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:16.5543518Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_cpp_serdes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:16.553949] 2025-12-04T09:50:23.5841883Z 2025-12-04T09:50:23.5842885Z export/test_cpp_serdes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_cpp_serdes_1.1_5b71872ad925eef5_.log 2025-12-04T09:50:23.5844058Z Running 0 items in this shard: 2025-12-04T09:50:23.5844272Z 2025-12-04T09:50:23.5844555Z Finished export/test_cpp_serdes 1/1 ... [2025-12-04 09:50:23.583768][2326.000241902], took 0.12min 2025-12-04T09:50:23.5914133Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_cpp_serdes/export.test_cpp_serdes-28f96f522b58f753.xml 2025-12-04T09:50:23.6666052Z Running inductor/test_control_flow 2/4 ... [2025-12-04 09:50:23.666206][2326.082678625] 2025-12-04T09:50:23.6666555Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:23.6669667Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_control_flow.py', '--shard-id=2', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:23.666528] 2025-12-04T09:50:31.1445072Z 2025-12-04T09:50:31.1446325Z inductor/test_control_flow 2/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_control_flow_2.4_594d15ae78e9e345_.log 2025-12-04T09:50:31.1447093Z Running 0 items in this shard: 2025-12-04T09:50:31.1447373Z 2025-12-04T09:50:31.1447884Z Finished inductor/test_control_flow 2/4 ... [2025-12-04 09:50:31.144130][2333.560603556], took 0.12min 2025-12-04T09:50:31.1519588Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-22c864d375691836.xml 2025-12-04T09:50:31.2235904Z Running test_sort_and_select 1/1 ... [2025-12-04 09:50:31.223220][2333.639691681] 2025-12-04T09:50:31.2236341Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:31.2239873Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_sort_and_select.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:31.223589] 2025-12-04T09:50:35.0953026Z 2025-12-04T09:50:35.0953904Z test_sort_and_select 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_sort_and_select_1.1_1cd4117a3aecc388_.log 2025-12-04T09:50:35.0954724Z Running 0 items in this shard: 2025-12-04T09:50:35.0954927Z 2025-12-04T09:50:35.0955250Z Finished test_sort_and_select 1/1 ... [2025-12-04 09:50:35.094965][2337.511438253], took 0.06min 2025-12-04T09:50:35.1030549Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_sort_and_select/test_sort_and_select-900006a153e58acf.xml 2025-12-04T09:50:35.1332030Z Running functorch/test_rearrange 1/1 ... [2025-12-04 09:50:35.132837][2337.549310778] 2025-12-04T09:50:35.1332507Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:35.1336436Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_rearrange.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:35.133177] 2025-12-04T09:50:38.4039586Z 2025-12-04T09:50:38.4040568Z functorch/test_rearrange 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_rearrange_1.1_a4effedcd0fd4055_.log 2025-12-04T09:50:38.4041341Z Running 0 items in this shard: 2025-12-04T09:50:38.4041527Z 2025-12-04T09:50:38.4041814Z Finished functorch/test_rearrange 1/1 ... [2025-12-04 09:50:38.403400][2340.819871367], took 0.05min 2025-12-04T09:50:38.4119674Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_rearrange/functorch.test_rearrange-91f40a7195b2f228.xml 2025-12-04T09:50:38.4371519Z Running test_package 1/1 ... [2025-12-04 09:50:38.436822][2340.85329547] 2025-12-04T09:50:38.4371960Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:38.4375657Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_package.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:38.437151] 2025-12-04T09:50:42.4590037Z 2025-12-04T09:50:42.4591044Z test_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_package_1.1_49fdbbc87a24207c_.log 2025-12-04T09:50:42.4603783Z Running 50 items in this shard: test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestAnalyze::test_trace_dependencies 2025-12-04T09:50:42.4616162Z 2025-12-04T09:50:42.4616433Z Finished test_package 1/1 ... [2025-12-04 09:50:42.458682][2344.87515551], took 0.07min 2025-12-04T09:50:42.4671264Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_package/test_package-4f9aecb7033550fa.xml 2025-12-04T09:50:42.5035149Z Running test_mkl_verbose 1/1 ... [2025-12-04 09:50:42.503171][2344.919644653] 2025-12-04T09:50:42.5035863Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:42.5039019Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mkl_verbose.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:42.503495] 2025-12-04T09:50:45.7241816Z 2025-12-04T09:50:45.7242681Z test_mkl_verbose 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkl_verbose_1.1_13126a41f6f75547_.log 2025-12-04T09:50:45.7243349Z Running 0 items in this shard: 2025-12-04T09:50:45.7243538Z 2025-12-04T09:50:45.7243789Z Finished test_mkl_verbose 1/1 ... [2025-12-04 09:50:45.723876][2348.140346093], took 0.05min 2025-12-04T09:50:45.7324320Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-7e432f71a73d0aad.xml 2025-12-04T09:50:45.7577000Z Running test_utils_config_module 1/1 ... [2025-12-04 09:50:45.757311][2348.173784036] 2025-12-04T09:50:45.7578114Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:45.7580801Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_utils_config_module.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:45.757635] 2025-12-04T09:50:49.0285016Z 2025-12-04T09:50:49.0285910Z test_utils_config_module 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_utils_config_module_1.1_619e18307e64a1f2_.log 2025-12-04T09:50:49.0286642Z Running 0 items in this shard: 2025-12-04T09:50:49.0286839Z 2025-12-04T09:50:49.0287192Z Finished test_utils_config_module 1/1 ... [2025-12-04 09:50:49.028175][2351.44464879], took 0.05min 2025-12-04T09:50:49.0368531Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-197d280bfbc0adc1.xml 2025-12-04T09:50:49.0677031Z Running test_hop_infra 1/1 ... [2025-12-04 09:50:49.067379][2351.483852777] 2025-12-04T09:50:49.0677553Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:49.0680703Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_hop_infra.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:49.067706] 2025-12-04T09:50:53.1396747Z 2025-12-04T09:50:53.1398346Z test_hop_infra 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_hop_infra_1.1_39540bfa16752ae6_.log 2025-12-04T09:50:53.1399621Z Running 0 items in this shard: 2025-12-04T09:50:53.1399994Z 2025-12-04T09:50:53.1400479Z Finished test_hop_infra 1/1 ... [2025-12-04 09:50:53.139260][2355.555732773], took 0.07min 2025-12-04T09:50:53.1481709Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_hop_infra/test_hop_infra-3a90bb4fa75fbe13.xml 2025-12-04T09:50:53.1819709Z Running test_appending_byte_serializer 1/1 ... [2025-12-04 09:50:53.181580][2355.598053595] 2025-12-04T09:50:53.1820690Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:53.1822508Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_appending_byte_serializer.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:53.181905] 2025-12-04T09:50:56.4023963Z 2025-12-04T09:50:56.4024760Z test_appending_byte_serializer 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_appending_byte_serializer_1.1_5c129e37d69d7dd8_.log 2025-12-04T09:50:56.4025570Z Running 0 items in this shard: 2025-12-04T09:50:56.4025763Z 2025-12-04T09:50:56.4026068Z Finished test_appending_byte_serializer 1/1 ... [2025-12-04 09:50:56.402109][2358.818582466], took 0.05min 2025-12-04T09:50:56.4113580Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-6e708074a2279afc.xml 2025-12-04T09:50:56.4437818Z Running test_ao_sparsity 1/1 ... [2025-12-04 09:50:56.443440][2358.859914143] 2025-12-04T09:50:56.4438248Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:50:56.4441715Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ao_sparsity.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:56.443792] 2025-12-04T09:51:00.2651040Z 2025-12-04T09:51:00.2651788Z test_ao_sparsity 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_ao_sparsity_1.1_a93db1bf7e7d6f1b_.log 2025-12-04T09:51:00.2652779Z Running 0 items in this shard: 2025-12-04T09:51:00.2652962Z 2025-12-04T09:51:00.2653211Z Finished test_ao_sparsity 1/1 ... [2025-12-04 09:51:00.264787][2362.681260785], took 0.06min 2025-12-04T09:51:00.2738520Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-51844dfab6f78c47.xml 2025-12-04T09:51:00.2997903Z Running test_extension_utils 1/1 ... [2025-12-04 09:51:00.299466][2362.715939681] 2025-12-04T09:51:00.2998347Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:00.3002492Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_extension_utils.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:00.299783] 2025-12-04T09:51:03.5207630Z 2025-12-04T09:51:03.5208620Z test_extension_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_extension_utils_1.1_82e1cd549caa8952_.log 2025-12-04T09:51:03.5209340Z Running 0 items in this shard: 2025-12-04T09:51:03.5209526Z 2025-12-04T09:51:03.5209797Z Finished test_extension_utils 1/1 ... [2025-12-04 09:51:03.520419][2365.936887992], took 0.05min 2025-12-04T09:51:03.5298313Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_extension_utils/test_extension_utils-9ca2aaa70629c80b.xml 2025-12-04T09:51:03.5563770Z Running nn/attention/test_fa4 1/1 ... [2025-12-04 09:51:03.555937][2365.972410722] 2025-12-04T09:51:03.5564211Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:03.5566856Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/attention/test_fa4.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:03.556254] 2025-12-04T09:51:07.3804734Z 2025-12-04T09:51:07.3806079Z nn/attention/test_fa4 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.attention.test_fa4_1.1_c635564b7b730bc9_.log 2025-12-04T09:51:07.3806998Z Running 0 items in this shard: 2025-12-04T09:51:07.3807286Z 2025-12-04T09:51:07.3807559Z Finished nn/attention/test_fa4 1/1 ... [2025-12-04 09:51:07.379999][2369.796470241], took 0.06min 2025-12-04T09:51:07.3898242Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.attention.test_fa4/nn.attention.test_fa4-d1bda12ba20d2196.xml 2025-12-04T09:51:07.4174705Z Running typing/test_python_operators 1/1 ... [2025-12-04 09:51:07.417054][2369.833527998] 2025-12-04T09:51:07.4175187Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:07.4177466Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'typing/test_python_operators.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:07.417356] 2025-12-04T09:51:11.1382677Z 2025-12-04T09:51:11.1383742Z typing/test_python_operators 1/1 was successful, full logs can be found in artifacts with path test/test-reports/typing.test_python_operators_1.1_042f5ad902ba34e2_.log 2025-12-04T09:51:11.1384530Z Running 0 items in this shard: 2025-12-04T09:51:11.1384719Z 2025-12-04T09:51:11.1385026Z Finished typing/test_python_operators 1/1 ... [2025-12-04 09:51:11.137920][2373.554393195], took 0.06min 2025-12-04T09:51:11.1475780Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/typing.test_python_operators/typing.test_python_operators-ba4b7d5e3b95dbfb.xml 2025-12-04T09:51:11.1710168Z Running torch_np/test_dtype 1/1 ... [2025-12-04 09:51:11.170594][2373.587067816] 2025-12-04T09:51:11.1710614Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:11.1712954Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_dtype.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:11.170898] 2025-12-04T09:51:14.4415521Z 2025-12-04T09:51:14.4416925Z torch_np/test_dtype 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_dtype_1.1_26a6994970880ab6_.log 2025-12-04T09:51:14.4417630Z Running 0 items in this shard: 2025-12-04T09:51:14.4417814Z 2025-12-04T09:51:14.4418073Z Finished torch_np/test_dtype 1/1 ... [2025-12-04 09:51:14.441250][2376.857722845], took 0.05min 2025-12-04T09:51:14.4512199Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_dtype/torch_np.test_dtype-58897eedc88355ca.xml 2025-12-04T09:51:14.4758854Z Running test_file_check 1/1 ... [2025-12-04 09:51:14.475473][2376.891946901] 2025-12-04T09:51:14.4759279Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:14.4763411Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_file_check.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:14.475810] 2025-12-04T09:51:17.6965539Z 2025-12-04T09:51:17.6966763Z test_file_check 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_file_check_1.1_a601f07c53907b26_.log 2025-12-04T09:51:17.6968113Z Running 0 items in this shard: 2025-12-04T09:51:17.6968451Z 2025-12-04T09:51:17.6968894Z Finished test_file_check 1/1 ... [2025-12-04 09:51:17.696052][2380.112522974], took 0.05min 2025-12-04T09:51:17.7065289Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_file_check/test_file_check-9ff5837f5f97472d.xml 2025-12-04T09:51:17.7331641Z Running profiler/test_kineto 1/1 ... [2025-12-04 09:51:17.732758][2380.149232539] 2025-12-04T09:51:17.7332155Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:17.7334390Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_kineto.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:17.733071] 2025-12-04T09:51:20.9529469Z 2025-12-04T09:51:20.9530667Z profiler/test_kineto 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_kineto_1.1_9a415b80c30847d8_.log 2025-12-04T09:51:20.9531397Z Running 0 items in this shard: 2025-12-04T09:51:20.9531588Z 2025-12-04T09:51:20.9531848Z Finished profiler/test_kineto 1/1 ... [2025-12-04 09:51:20.952607][2383.369080127], took 0.05min 2025-12-04T09:51:20.9628605Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_kineto/profiler.test_kineto-81a44d1732bd85bc.xml 2025-12-04T09:51:20.9876968Z Running functorch/test_ac_knapsack 1/1 ... [2025-12-04 09:51:20.987378][2383.403851322] 2025-12-04T09:51:20.9877445Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:20.9880959Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ac_knapsack.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:20.987705] 2025-12-04T09:51:24.3579571Z 2025-12-04T09:51:24.3580563Z functorch/test_ac_knapsack 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ac_knapsack_1.1_63c0b4d8b55c8c85_.log 2025-12-04T09:51:24.3581334Z Running 0 items in this shard: 2025-12-04T09:51:24.3581542Z 2025-12-04T09:51:24.3581835Z Finished functorch/test_ac_knapsack 1/1 ... [2025-12-04 09:51:24.357672][2386.774143434], took 0.06min 2025-12-04T09:51:24.3684833Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ac_knapsack/functorch.test_ac_knapsack-56b46ef5e2d4cc8d.xml 2025-12-04T09:51:24.3949734Z Running torch_np/test_nep50_examples 1/1 ... [2025-12-04 09:51:24.394602][2386.811074956] 2025-12-04T09:51:24.3950365Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:24.3952983Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_nep50_examples.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:24.394917] 2025-12-04T09:51:30.2187271Z 2025-12-04T09:51:30.2188090Z torch_np/test_nep50_examples 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_nep50_examples_1.1_d88ffefdf604b471_.log 2025-12-04T09:51:30.2189030Z Running 0 items in this shard: 2025-12-04T09:51:30.2189222Z 2025-12-04T09:51:30.2189520Z Finished torch_np/test_nep50_examples 1/1 ... [2025-12-04 09:51:30.218384][2392.634857179], took 0.10min 2025-12-04T09:51:30.2291070Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_nep50_examples/torch_np.test_nep50_examples-a5a9215260d58c1d.xml 2025-12-04T09:51:30.2549336Z Running test_torch 1/1 ... [2025-12-04 09:51:30.254580][2392.671053099] 2025-12-04T09:51:30.2549752Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:30.2552998Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_torch.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:30.254915] 2025-12-04T09:51:56.6852349Z 2025-12-04T09:51:56.6853067Z PRINTING LOG FILE of test_torch 1/1 (test/test-reports/test_torch_1.1_a0ae95f961ed9b78_.log) 2025-12-04T09:51:56.6855600Z Test results will be stored in test-reports/python-pytest/test_torch/test_torch-3d54dab073791e89.xml 2025-12-04T09:51:56.6856410Z ============================= test session starts ============================== 2025-12-04T09:51:56.6857129Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:51:56.6857622Z cachedir: .pytest_cache 2025-12-04T09:51:56.6858196Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:51:56.6859028Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:51:56.6859323Z configfile: pytest.ini 2025-12-04T09:51:56.6859963Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:51:56.6860823Z collecting ... collected 976 items 2025-12-04T09:51:56.6861232Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:51:56.6891286Z Running 150 items in this shard: test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_index_add_correctness, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_qengine, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup, test/test_torch.py::TestTorch::test_tensoriterator_output_setup 2025-12-04T09:51:56.6920390Z 2025-12-04T09:51:56.6920651Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.7590s] [ 0%] 2025-12-04T09:51:56.6921573Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 1%] 2025-12-04T09:51:56.6922811Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6923746Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4076s] [ 2%] 2025-12-04T09:51:56.6924360Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3824s] [ 2%] 2025-12-04T09:51:56.6924928Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3927s] [ 2%] 2025-12-04T09:51:56.6925477Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3795s] [ 2%] 2025-12-04T09:51:56.6926033Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3898s] [ 2%] 2025-12-04T09:51:56.6926638Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4054s] [ 2%] 2025-12-04T09:51:56.6927197Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3975s] [ 2%] 2025-12-04T09:51:56.6927914Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4065s] [ 2%] 2025-12-04T09:51:56.6928471Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3780s] [ 2%] 2025-12-04T09:51:56.6929028Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3835s] [ 2%] 2025-12-04T09:51:56.6929642Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3846s] [ 2%] 2025-12-04T09:51:56.6930195Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3791s] [ 2%] 2025-12-04T09:51:56.6930758Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3989s] [ 2%] 2025-12-04T09:51:56.6931311Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3801s] [ 2%] 2025-12-04T09:51:56.6931863Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3889s] [ 2%] 2025-12-04T09:51:56.6932417Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3969s] [ 2%] 2025-12-04T09:51:56.6932970Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3965s] [ 2%] 2025-12-04T09:51:56.6933522Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3688s] [ 2%] 2025-12-04T09:51:56.6934068Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3876s] [ 2%] 2025-12-04T09:51:56.6934627Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4044s] [ 2%] 2025-12-04T09:51:56.6935185Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3877s] [ 2%] 2025-12-04T09:51:56.6936056Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3867s] [ 2%] 2025-12-04T09:51:56.6936607Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4053s] [ 2%] 2025-12-04T09:51:56.6937166Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4007s] [ 2%] 2025-12-04T09:51:56.6937722Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3763s] [ 2%] 2025-12-04T09:51:56.6938267Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3951s] [ 2%] 2025-12-04T09:51:56.6938817Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3942s] [ 2%] 2025-12-04T09:51:56.6939371Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3918s] [ 2%] 2025-12-04T09:51:56.6939979Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3933s] [ 2%] 2025-12-04T09:51:56.6940524Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3719s] [ 2%] 2025-12-04T09:51:56.6941165Z test_torch.py::TestTorch::test_index_add_correctness FAILED [0.2887s] [ 2%] 2025-12-04T09:51:56.6941721Z test_torch.py::TestTorch::test_index_add_correctness FAILED [0.3144s] [ 2%] 2025-12-04T09:51:56.6942263Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3996s] [ 2%] 2025-12-04T09:51:56.6942823Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4149s] [ 2%] 2025-12-04T09:51:56.6943377Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4362s] [ 2%] 2025-12-04T09:51:56.6943930Z test_torch.py::TestTorch::test_index_add_correctness FAILED [0.1627s] [ 2%] 2025-12-04T09:51:56.6944471Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4121s] [ 2%] 2025-12-04T09:51:56.6945027Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4048s] [ 2%] 2025-12-04T09:51:56.6945643Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4186s] [ 2%] 2025-12-04T09:51:56.6946193Z test_torch.py::TestTorch::test_index_add_correctness FAILED [0.2858s] [ 2%] 2025-12-04T09:51:56.6946749Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4003s] [ 2%] 2025-12-04T09:51:56.6947304Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4196s] [ 2%] 2025-12-04T09:51:56.6947952Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4108s] [ 2%] 2025-12-04T09:51:56.6948502Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3819s] [ 2%] 2025-12-04T09:51:56.6949052Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4160s] [ 2%] 2025-12-04T09:51:56.6949607Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.3817s] [ 2%] 2025-12-04T09:51:56.6950158Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4146s] [ 2%] 2025-12-04T09:51:56.6950765Z test_torch.py::TestTorch::test_index_add_correctness PASSED [0.4389s] [ 2%] 2025-12-04T09:51:56.6951323Z test_torch.py::TestTorch::test_index_add_correctness FAILED [0.3173s] [ 2%] 2025-12-04T09:51:56.6952183Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6953327Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6954484Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6955626Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0008s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6956774Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6958105Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6959255Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6960403Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6961679Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6963153Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6964397Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6965732Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6966901Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6968121Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6969262Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6970527Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6971689Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6972877Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6974126Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6975280Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6976468Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6977622Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6978773Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6979976Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6981124Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6982273Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6983434Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6984589Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6985738Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6986884Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6988035Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6989236Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6990440Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6991597Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6992742Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6993904Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6995108Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6996255Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6997494Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6998647Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.6999797Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7001035Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7002194Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7003349Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7013580Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7014803Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7016078Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7017244Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7018394Z test_torch.py::TestTorch::test_qengine SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7019628Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7020936Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7022246Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7023646Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7024965Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7026275Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7027575Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7028880Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7030237Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7031579Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7032884Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0010s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7034177Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7035892Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7037212Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7038523Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7039823Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7041117Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7042432Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7043730Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7045036Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7046340Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7047698Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7049120Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7050476Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7051775Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7053076Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7054373Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7055737Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7057098Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7058400Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7059702Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7061053Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7062357Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7063664Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7064957Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7066256Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7067561Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7068857Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7070216Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7071515Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7072823Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7074169Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7075479Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0007s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7076783Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7078078Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7079368Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7080702Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7082036Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7083328Z test_torch.py::TestTorch::test_tensoriterator_output_setup SKIPPED [0.0006s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:51:56.7084026Z 2025-12-04T09:51:56.7084162Z =================================== FAILURES =================================== 2025-12-04T09:51:56.7084631Z _____________________ TestTorch.test_index_add_correctness _____________________ 2025-12-04T09:51:56.7085049Z Traceback (most recent call last): 2025-12-04T09:51:56.7085561Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6717, in test_index_add_correctness 2025-12-04T09:51:56.7086089Z helper(dim, dtype, device, size, size) 2025-12-04T09:51:56.7086526Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6708, in helper 2025-12-04T09:51:56.7087025Z self.assertEqual(out, ref_out, atol=1e-2, rtol=1e-2) 2025-12-04T09:51:56.7087769Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:51:56.7088472Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:51:56.7088879Z AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7089111Z 2025-12-04T09:51:56.7089219Z Mismatched elements: 1 / 262144 (0.0%) 2025-12-04T09:51:56.7089657Z Greatest absolute difference: 0.03125 at index (0, 163, 211) (up to 0.01 allowed) 2025-12-04T09:51:56.7090250Z Greatest relative difference: 0.0150146484375 at index (0, 163, 211) (up to 0.01 allowed) 2025-12-04T09:51:56.7090622Z 2025-12-04T09:51:56.7090809Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7091520Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7092045Z 2025-12-04T09:51:56.7092277Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7092795Z _____________________ TestTorch.test_index_add_correctness _____________________ 2025-12-04T09:51:56.7093205Z Traceback (most recent call last): 2025-12-04T09:51:56.7093700Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6717, in test_index_add_correctness 2025-12-04T09:51:56.7094209Z helper(dim, dtype, device, size, size) 2025-12-04T09:51:56.7094647Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6708, in helper 2025-12-04T09:51:56.7095139Z self.assertEqual(out, ref_out, atol=1e-2, rtol=1e-2) 2025-12-04T09:51:56.7095865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:51:56.7096565Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:51:56.7096969Z AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7097190Z 2025-12-04T09:51:56.7097307Z Mismatched elements: 1 / 327680 (0.0%) 2025-12-04T09:51:56.7097734Z Greatest absolute difference: 0.03125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7098329Z Greatest relative difference: 0.01495361328125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7098707Z 2025-12-04T09:51:56.7098887Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7099624Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7100163Z 2025-12-04T09:51:56.7100429Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7100960Z _____________________ TestTorch.test_index_add_correctness _____________________ 2025-12-04T09:51:56.7101372Z Traceback (most recent call last): 2025-12-04T09:51:56.7101867Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6717, in test_index_add_correctness 2025-12-04T09:51:56.7102443Z helper(dim, dtype, device, size, size) 2025-12-04T09:51:56.7102886Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6708, in helper 2025-12-04T09:51:56.7103373Z self.assertEqual(out, ref_out, atol=1e-2, rtol=1e-2) 2025-12-04T09:51:56.7104037Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:51:56.7104733Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:51:56.7105170Z AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7105389Z 2025-12-04T09:51:56.7105505Z Mismatched elements: 1 / 327680 (0.0%) 2025-12-04T09:51:56.7105929Z Greatest absolute difference: 0.03125 at index (4, 231, 225) (up to 0.01 allowed) 2025-12-04T09:51:56.7106536Z Greatest relative difference: 0.01495361328125 at index (4, 231, 225) (up to 0.01 allowed) 2025-12-04T09:51:56.7106903Z 2025-12-04T09:51:56.7107091Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7107807Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7108331Z 2025-12-04T09:51:56.7108553Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7109076Z _____________________ TestTorch.test_index_add_correctness _____________________ 2025-12-04T09:51:56.7109482Z Traceback (most recent call last): 2025-12-04T09:51:56.7109970Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6717, in test_index_add_correctness 2025-12-04T09:51:56.7110497Z helper(dim, dtype, device, size, size) 2025-12-04T09:51:56.7110935Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6708, in helper 2025-12-04T09:51:56.7111426Z self.assertEqual(out, ref_out, atol=1e-2, rtol=1e-2) 2025-12-04T09:51:56.7112086Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:51:56.7112779Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:51:56.7113178Z AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7113398Z 2025-12-04T09:51:56.7113520Z Mismatched elements: 1 / 327680 (0.0%) 2025-12-04T09:51:56.7113943Z Greatest absolute difference: 0.03125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7114535Z Greatest relative difference: 0.01495361328125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7114895Z 2025-12-04T09:51:56.7115089Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7115798Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7116386Z 2025-12-04T09:51:56.7116607Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7117130Z _____________________ TestTorch.test_index_add_correctness _____________________ 2025-12-04T09:51:56.7117542Z Traceback (most recent call last): 2025-12-04T09:51:56.7118034Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6717, in test_index_add_correctness 2025-12-04T09:51:56.7118551Z helper(dim, dtype, device, size, size) 2025-12-04T09:51:56.7118992Z File "/var/lib/jenkins/workspace/test/test_torch.py", line 6708, in helper 2025-12-04T09:51:56.7119472Z self.assertEqual(out, ref_out, atol=1e-2, rtol=1e-2) 2025-12-04T09:51:56.7120135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:51:56.7120886Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:51:56.7121286Z AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7121512Z 2025-12-04T09:51:56.7121616Z Mismatched elements: 1 / 327680 (0.0%) 2025-12-04T09:51:56.7122049Z Greatest absolute difference: 0.03125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7122696Z Greatest relative difference: 0.01495361328125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7123058Z 2025-12-04T09:51:56.7123248Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7123939Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7124464Z 2025-12-04T09:51:56.7124683Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7125494Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_torch/test_torch-3d54dab073791e89.xml - 2025-12-04T09:51:56.7126192Z =========================== short test summary info ============================ 2025-12-04T09:51:56.7126796Z FAILED [0.2887s] test_torch.py::TestTorch::test_index_add_correctness - AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7127255Z 2025-12-04T09:51:56.7127416Z Mismatched elements: 1 / 262144 (0.0%) 2025-12-04T09:51:56.7127852Z Greatest absolute difference: 0.03125 at index (0, 163, 211) (up to 0.01 allowed) 2025-12-04T09:51:56.7128437Z Greatest relative difference: 0.0150146484375 at index (0, 163, 211) (up to 0.01 allowed) 2025-12-04T09:51:56.7128802Z 2025-12-04T09:51:56.7128979Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7129704Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7130256Z 2025-12-04T09:51:56.7130480Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7131154Z FAILED [0.3144s] test_torch.py::TestTorch::test_index_add_correctness - AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7131610Z 2025-12-04T09:51:56.7131719Z Mismatched elements: 1 / 327680 (0.0%) 2025-12-04T09:51:56.7132158Z Greatest absolute difference: 0.03125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7132747Z Greatest relative difference: 0.01495361328125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7133107Z 2025-12-04T09:51:56.7133288Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7133991Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7134525Z 2025-12-04T09:51:56.7134749Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7135666Z FAILED [0.1627s] test_torch.py::TestTorch::test_index_add_correctness - AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7136136Z 2025-12-04T09:51:56.7136242Z Mismatched elements: 1 / 327680 (0.0%) 2025-12-04T09:51:56.7136762Z Greatest absolute difference: 0.03125 at index (4, 231, 225) (up to 0.01 allowed) 2025-12-04T09:51:56.7137355Z Greatest relative difference: 0.01495361328125 at index (4, 231, 225) (up to 0.01 allowed) 2025-12-04T09:51:56.7137721Z 2025-12-04T09:51:56.7137913Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7138609Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7139137Z 2025-12-04T09:51:56.7139359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7140039Z FAILED [0.2858s] test_torch.py::TestTorch::test_index_add_correctness - AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7140495Z 2025-12-04T09:51:56.7140609Z Mismatched elements: 1 / 327680 (0.0%) 2025-12-04T09:51:56.7141094Z Greatest absolute difference: 0.03125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7141693Z Greatest relative difference: 0.01495361328125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7142064Z 2025-12-04T09:51:56.7142251Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7143010Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7143529Z 2025-12-04T09:51:56.7143752Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7144433Z FAILED [0.3173s] test_torch.py::TestTorch::test_index_add_correctness - AssertionError: Tensor-likes are not close! 2025-12-04T09:51:56.7144895Z 2025-12-04T09:51:56.7145001Z Mismatched elements: 1 / 327680 (0.0%) 2025-12-04T09:51:56.7145497Z Greatest absolute difference: 0.03125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7146090Z Greatest relative difference: 0.01495361328125 at index (1, 17, 168) (up to 0.01 allowed) 2025-12-04T09:51:56.7146475Z 2025-12-04T09:51:56.7146658Z To execute this test, run the following from the base repo dir: 2025-12-04T09:51:56.7147359Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_torch.py TestTorch.test_index_add_correctness 2025-12-04T09:51:56.7147878Z 2025-12-04T09:51:56.7148105Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:51:56.7148591Z ================== 5 failed, 45 passed, 100 skipped in 21.33s ================== 2025-12-04T09:51:56.7148870Z 2025-12-04T09:51:56.7149174Z FINISHED PRINTING LOG FILE of test_torch 1/1 (test/test-reports/test_torch_1.1_a0ae95f961ed9b78_.log) 2025-12-04T09:51:56.7149576Z 2025-12-04T09:51:56.7149796Z Finished test_torch 1/1 ... [2025-12-04 09:51:56.685431][2419.101902592], took 0.44min 2025-12-04T09:51:56.7150613Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_torch/test_torch-3d54dab073791e89.xml 2025-12-04T09:51:57.5225612Z Uploading logs for 57118183167 to S3 2025-12-04T09:51:57.6150346Z Uploading artifacts took 0.85 seconds 2025-12-04T09:51:57.6150677Z test_torch 1/1 failed! 2025-12-04T09:51:57.6154973Z Running xpu/test_gemm 1/1 ... [2025-12-04 09:51:57.615195][2420.03166576] 2025-12-04T09:51:57.6155383Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:51:57.6160015Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'xpu/test_gemm.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:51:57.615643] 2025-12-04T09:52:01.5374574Z 2025-12-04T09:52:01.5375414Z xpu/test_gemm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/xpu.test_gemm_1.1_105887c98b5f57e5_.log 2025-12-04T09:52:01.5376079Z Running 0 items in this shard: 2025-12-04T09:52:01.5376283Z 2025-12-04T09:52:01.5376510Z Finished xpu/test_gemm 1/1 ... [2025-12-04 09:52:01.537172][2423.953644223], took 0.07min 2025-12-04T09:52:01.5482910Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-b46323f728df1f56.xml 2025-12-04T09:52:01.5722735Z Running test_binary_ufuncs 1/1 ... [2025-12-04 09:52:01.571855][2423.988329142] 2025-12-04T09:52:01.5723356Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:52:01.5725330Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_binary_ufuncs.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:52:01.572191] 2025-12-04T09:52:27.7285802Z 2025-12-04T09:52:27.7286539Z test_binary_ufuncs 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_binary_ufuncs_1.1_002f41f19aca627a_.log 2025-12-04T09:52:27.7287590Z Running 0 items in this shard: 2025-12-04T09:52:27.7287778Z 2025-12-04T09:52:27.7288037Z Finished test_binary_ufuncs 1/1 ... [2025-12-04 09:52:27.728268][2450.144741757], took 0.44min 2025-12-04T09:52:27.7395487Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_binary_ufuncs/test_binary_ufuncs-3281db6a17fd8e68.xml 2025-12-04T09:52:27.8032572Z Running test_modules 2/4 ... [2025-12-04 09:52:27.802933][2450.219403649] 2025-12-04T09:52:27.8033010Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:52:27.8036646Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_modules.py', '--shard-id=2', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:52:27.803276] 2025-12-04T09:52:38.6869262Z 2025-12-04T09:52:38.6870119Z test_modules 2/4 was successful, full logs can be found in artifacts with path test/test-reports/test_modules_2.4_328d1f3d3b0b68db_.log 2025-12-04T09:52:38.6888348Z Running 50 items in this shard: test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32 2025-12-04T09:52:38.6905414Z 2025-12-04T09:52:38.6905645Z Finished test_modules 2/4 ... [2025-12-04 09:52:38.686661][2461.103133477], took 0.18min 2025-12-04T09:52:38.6980492Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_modules/test_modules-0165bb40c77358d1.xml 2025-12-04T09:52:38.7703244Z Running torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-12-04 09:52:38.769892][2461.186364766] 2025-12-04T09:52:38.7703923Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:52:38.7706518Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/linalg/test_linalg.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:52:38.770244] 2025-12-04T09:52:42.4920272Z 2025-12-04T09:52:42.4921425Z torch_np/numpy_tests/linalg/test_linalg 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_e5ceed266a7995b6_.log 2025-12-04T09:52:42.4922338Z Running 0 items in this shard: 2025-12-04T09:52:42.4922534Z 2025-12-04T09:52:42.4922872Z Finished torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-12-04 09:52:42.491701][2464.908174483], took 0.06min 2025-12-04T09:52:42.5037676Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-696665a6c90efdc4.xml 2025-12-04T09:52:42.5305768Z Running torch_np/numpy_tests/core/test_dtype 1/1 ... [2025-12-04 09:52:42.530201][2464.946674005] 2025-12-04T09:52:42.5306393Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:52:42.5309546Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_dtype.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:52:42.530555] 2025-12-04T09:52:45.9016072Z 2025-12-04T09:52:45.9016947Z torch_np/numpy_tests/core/test_dtype 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_dtype_1.1_12269e77e29080a2_.log 2025-12-04T09:52:45.9018003Z Running 0 items in this shard: 2025-12-04T09:52:45.9018187Z 2025-12-04T09:52:45.9018514Z Finished torch_np/numpy_tests/core/test_dtype 1/1 ... [2025-12-04 09:52:45.901331][2468.317804361], took 0.06min 2025-12-04T09:52:45.9128211Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-f968ba80b11f10a3.xml 2025-12-04T09:52:45.9436334Z Running lazy/test_debug_util 1/1 ... [2025-12-04 09:52:45.943213][2468.3596865] 2025-12-04T09:52:45.9436790Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:52:45.9441695Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_debug_util.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:52:45.943583] 2025-12-04T09:52:49.1642795Z 2025-12-04T09:52:49.1643788Z lazy/test_debug_util 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_debug_util_1.1_ce36b8dbe1e238fe_.log 2025-12-04T09:52:49.1644519Z Running 0 items in this shard: 2025-12-04T09:52:49.1644707Z 2025-12-04T09:52:49.1644960Z Finished lazy/test_debug_util 1/1 ... [2025-12-04 09:52:49.163828][2471.580301933], took 0.05min 2025-12-04T09:52:49.1756992Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-953e7ce7fc6caa5f.xml 2025-12-04T09:52:49.2020347Z Running nn/test_load_state_dict 1/1 ... [2025-12-04 09:52:49.201687][2471.618160575] 2025-12-04T09:52:49.2020850Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:52:49.2025032Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_load_state_dict.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:52:49.202098] 2025-12-04T09:52:52.8232338Z 2025-12-04T09:52:52.8233212Z nn/test_load_state_dict 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_load_state_dict_1.1_6514e2b2f6071409_.log 2025-12-04T09:52:52.8233941Z Running 0 items in this shard: 2025-12-04T09:52:52.8234126Z 2025-12-04T09:52:52.8234396Z Finished nn/test_load_state_dict 1/1 ... [2025-12-04 09:52:52.822958][2475.239431328], took 0.06min 2025-12-04T09:52:52.8349842Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-48f47027a0649f8b.xml 2025-12-04T09:52:52.8594792Z Running test_shape_ops 1/1 ... [2025-12-04 09:52:52.859008][2475.27548219] 2025-12-04T09:52:52.8595361Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:52:52.8597851Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_shape_ops.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:52:52.859357] 2025-12-04T09:52:56.7315475Z 2025-12-04T09:52:56.7316624Z test_shape_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_shape_ops_1.1_42da27b0199bee14_.log 2025-12-04T09:52:56.7317284Z Running 0 items in this shard: 2025-12-04T09:52:56.7317468Z 2025-12-04T09:52:56.7317704Z Finished test_shape_ops 1/1 ... [2025-12-04 09:52:56.730972][2479.147441357], took 0.06min 2025-12-04T09:52:56.7435845Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_shape_ops/test_shape_ops-85cca46ea4083bfa.xml 2025-12-04T09:52:56.7712926Z Running profiler/test_memory_profiler 1/1 ... [2025-12-04 09:52:56.770932][2479.187406055] 2025-12-04T09:52:56.7713460Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:52:56.7716714Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_memory_profiler.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:52:56.771278] 2025-12-04T09:53:00.6426448Z 2025-12-04T09:53:00.6429082Z profiler/test_memory_profiler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_memory_profiler_1.1_6e75fd2681e33996_.log 2025-12-04T09:53:00.6451532Z Running 50 items in this shard: test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none 2025-12-04T09:53:00.6472804Z 2025-12-04T09:53:00.6473109Z Finished profiler/test_memory_profiler 1/1 ... [2025-12-04 09:53:00.642382][2483.058854613], took 0.06min 2025-12-04T09:53:00.6547196Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-b02e0fce99e01635.xml 2025-12-04T09:53:00.7143694Z Running test_indexing 1/1 ... [2025-12-04 09:53:00.714003][2483.13047612] 2025-12-04T09:53:00.7144262Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:00.7147697Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_indexing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:00.714370] 2025-12-04T09:53:04.7364320Z 2025-12-04T09:53:04.7365038Z test_indexing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_indexing_1.1_dca170216ef563f6_.log 2025-12-04T09:53:04.7365938Z Running 0 items in this shard: 2025-12-04T09:53:04.7366125Z 2025-12-04T09:53:04.7366360Z Finished test_indexing 1/1 ... [2025-12-04 09:53:04.736041][2487.152509176], took 0.07min 2025-12-04T09:53:04.7489596Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_indexing/test_indexing-c851479f46f8a70b.xml 2025-12-04T09:53:04.7702309Z Running test_type_info 1/1 ... [2025-12-04 09:53:04.769883][2487.186356992] 2025-12-04T09:53:04.7702736Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:04.7705968Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_type_info.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:04.770227] 2025-12-04T09:53:07.9905772Z 2025-12-04T09:53:07.9906768Z test_type_info 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_type_info_1.1_586f6a1b874ca23b_.log 2025-12-04T09:53:07.9907597Z Running 0 items in this shard: 2025-12-04T09:53:07.9907783Z 2025-12-04T09:53:07.9908018Z Finished test_type_info 1/1 ... [2025-12-04 09:53:07.990262][2490.406735602], took 0.05min 2025-12-04T09:53:08.0029366Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_type_info/test_type_info-1c1348d8e4ba037e.xml 2025-12-04T09:53:08.0284252Z Running functorch/test_aotdispatch 1/1 ... [2025-12-04 09:53:08.028090][2490.444562484] 2025-12-04T09:53:08.0284825Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:08.0288068Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_aotdispatch.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:08.028428] 2025-12-04T09:53:14.4044792Z 2025-12-04T09:53:14.4046170Z functorch/test_aotdispatch 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_aotdispatch_1.1_e6cf91daf0dd7020_.log 2025-12-04T09:53:14.4047467Z Running 0 items in this shard: 2025-12-04T09:53:14.4047726Z 2025-12-04T09:53:14.4048187Z Finished functorch/test_aotdispatch 1/1 ... [2025-12-04 09:53:14.404107][2496.820579903], took 0.11min 2025-12-04T09:53:14.4170894Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_aotdispatch/functorch.test_aotdispatch-0f6db4e497a4ab5e.xml 2025-12-04T09:53:14.4713562Z Running test_scatter_gather_ops 1/1 ... [2025-12-04 09:53:14.470987][2496.887459431] 2025-12-04T09:53:14.4714224Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:14.4718974Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_scatter_gather_ops.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:14.471356] 2025-12-04T09:53:18.3434032Z 2025-12-04T09:53:18.3435077Z test_scatter_gather_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_scatter_gather_ops_1.1_7a88c857d981a5d2_.log 2025-12-04T09:53:18.3436554Z Running 0 items in this shard: 2025-12-04T09:53:18.3436828Z 2025-12-04T09:53:18.3437228Z Finished test_scatter_gather_ops 1/1 ... [2025-12-04 09:53:18.342984][2500.759454605], took 0.06min 2025-12-04T09:53:18.3562637Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_scatter_gather_ops/test_scatter_gather_ops-96cbf94077b2f70a.xml 2025-12-04T09:53:18.3783901Z Running test_cuda_multigpu 1/1 ... [2025-12-04 09:53:18.378037][2500.794510784] 2025-12-04T09:53:18.3784363Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:18.3787886Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_cuda_multigpu.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:18.378388] 2025-12-04T09:53:29.8168528Z 2025-12-04T09:53:29.8169281Z PRINTING LOG FILE of test_cuda_multigpu 1/1 (test/test-reports/test_cuda_multigpu_1.1_a23ea26f9721e68d_.log) 2025-12-04T09:53:29.8170357Z Test results will be stored in test-reports/python-pytest/test_cuda_multigpu/test_cuda_multigpu-4895f15a6c4e5281.xml 2025-12-04T09:53:29.8171183Z ============================= test session starts ============================== 2025-12-04T09:53:29.8172005Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:53:29.8172728Z cachedir: .pytest_cache 2025-12-04T09:53:29.8173419Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:53:29.8174604Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:53:29.8174994Z configfile: pytest.ini 2025-12-04T09:53:29.8175664Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:53:29.8176531Z collecting ... collected 61 items 2025-12-04T09:53:29.8177152Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:53:29.8191778Z Running 50 items in this shard: test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection, test/test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection 2025-12-04T09:53:29.8206374Z 2025-12-04T09:53:29.8206682Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.2549s] [ 2%] 2025-12-04T09:53:29.8207475Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.1227s] [ 2%] 2025-12-04T09:53:29.8208166Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.2025s] [ 2%] 2025-12-04T09:53:29.8208880Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.1022s] [ 2%] 2025-12-04T09:53:29.8209724Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.2029s] [ 2%] 2025-12-04T09:53:29.8210527Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.1019s] [ 2%] 2025-12-04T09:53:29.8211320Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.2020s] [ 2%] 2025-12-04T09:53:29.8212119Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.1023s] [ 2%] 2025-12-04T09:53:29.8212911Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.2036s] [ 2%] 2025-12-04T09:53:29.8213717Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.1021s] [ 2%] 2025-12-04T09:53:29.8214520Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1994s] [ 2%] 2025-12-04T09:53:29.8215313Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.1003s] [ 2%] 2025-12-04T09:53:29.8216115Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1977s] [ 2%] 2025-12-04T09:53:29.8216915Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0997s] [ 2%] 2025-12-04T09:53:29.8217715Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1978s] [ 2%] 2025-12-04T09:53:29.8218501Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0998s] [ 2%] 2025-12-04T09:53:29.8232313Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1984s] [ 2%] 2025-12-04T09:53:29.8233023Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0997s] [ 2%] 2025-12-04T09:53:29.8233716Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1971s] [ 2%] 2025-12-04T09:53:29.8234400Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0997s] [ 2%] 2025-12-04T09:53:29.8235096Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1978s] [ 2%] 2025-12-04T09:53:29.8236097Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0999s] [ 2%] 2025-12-04T09:53:29.8236928Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1975s] [ 2%] 2025-12-04T09:53:29.8237617Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0998s] [ 2%] 2025-12-04T09:53:29.8238304Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1986s] [ 2%] 2025-12-04T09:53:29.8238992Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0995s] [ 2%] 2025-12-04T09:53:29.8239667Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1973s] [ 2%] 2025-12-04T09:53:29.8240348Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0999s] [ 2%] 2025-12-04T09:53:29.8241031Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1973s] [ 2%] 2025-12-04T09:53:29.8241789Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0998s] [ 2%] 2025-12-04T09:53:29.8242464Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1975s] [ 2%] 2025-12-04T09:53:29.8243155Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0997s] [ 2%] 2025-12-04T09:53:29.8243919Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1983s] [ 2%] 2025-12-04T09:53:29.8244606Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0999s] [ 2%] 2025-12-04T09:53:29.8245285Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1976s] [ 2%] 2025-12-04T09:53:29.8245999Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0996s] [ 2%] 2025-12-04T09:53:29.8246770Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1975s] [ 2%] 2025-12-04T09:53:29.8247549Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0999s] [ 2%] 2025-12-04T09:53:29.8248232Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1977s] [ 2%] 2025-12-04T09:53:29.8248915Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.1000s] [ 2%] 2025-12-04T09:53:29.8249597Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1986s] [ 2%] 2025-12-04T09:53:29.8250275Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0999s] [ 2%] 2025-12-04T09:53:29.8250965Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1973s] [ 2%] 2025-12-04T09:53:29.8251643Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0997s] [ 2%] 2025-12-04T09:53:29.8252321Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1975s] [ 2%] 2025-12-04T09:53:29.8253006Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0999s] [ 2%] 2025-12-04T09:53:29.8253689Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1976s] [ 2%] 2025-12-04T09:53:29.8254369Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0998s] [ 2%] 2025-12-04T09:53:29.8255104Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection PASSED [0.1986s] [ 2%] 2025-12-04T09:53:29.8255779Z test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection FAILED [0.0998s] [ 2%] 2025-12-04T09:53:29.8256175Z 2025-12-04T09:53:29.8256299Z =================================== FAILURES =================================== 2025-12-04T09:53:29.8256748Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8257179Z Traceback (most recent call last): 2025-12-04T09:53:29.8257732Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8258342Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8258894Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8259402Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8259916Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8260426Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8260771Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8260985Z 2025-12-04T09:53:29.8261168Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8261952Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8262561Z 2025-12-04T09:53:29.8262789Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8263347Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8265064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8266603Z with policy(): 2025-12-04T09:53:29.8266962Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8267393Z Traceback (most recent call last): 2025-12-04T09:53:29.8267937Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8268619Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8269117Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8269625Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8270138Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8270647Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8270991Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8271202Z 2025-12-04T09:53:29.8271383Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8272154Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8272748Z 2025-12-04T09:53:29.8272973Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8273471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8275108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8276634Z with policy(): 2025-12-04T09:53:29.8276944Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8278569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8280102Z with policy(): 2025-12-04T09:53:29.8280450Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8280926Z Traceback (most recent call last): 2025-12-04T09:53:29.8281475Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8282086Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8282589Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8283094Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8283607Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8284111Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8284451Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8284662Z 2025-12-04T09:53:29.8284870Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8285725Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8286327Z 2025-12-04T09:53:29.8286549Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8287050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8288780Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8290333Z with policy(): 2025-12-04T09:53:29.8290643Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8292331Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8293872Z with policy(): 2025-12-04T09:53:29.8294233Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8294662Z Traceback (most recent call last): 2025-12-04T09:53:29.8295209Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8295824Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8296315Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8296821Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8297340Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8297843Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8298182Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8298395Z 2025-12-04T09:53:29.8298575Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8299361Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8299957Z 2025-12-04T09:53:29.8300182Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8300678Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8302370Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8303911Z with policy(): 2025-12-04T09:53:29.8304217Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8305900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8307425Z with policy(): 2025-12-04T09:53:29.8307777Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8308254Z Traceback (most recent call last): 2025-12-04T09:53:29.8308799Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8309425Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8309916Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8310422Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8310996Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8311501Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8311843Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8312053Z 2025-12-04T09:53:29.8312233Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8313118Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8313864Z 2025-12-04T09:53:29.8314090Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8314595Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8316294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8317825Z with policy(): 2025-12-04T09:53:29.8318132Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8319772Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8321306Z with policy(): 2025-12-04T09:53:29.8321659Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8322089Z Traceback (most recent call last): 2025-12-04T09:53:29.8322638Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8323249Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8323743Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8324248Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8324758Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8325266Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8325605Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8325821Z 2025-12-04T09:53:29.8326057Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8326835Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8327533Z 2025-12-04T09:53:29.8327759Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8328262Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8329914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8331524Z with policy(): 2025-12-04T09:53:29.8331838Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8333527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8335088Z with policy(): 2025-12-04T09:53:29.8335669Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8336107Z Traceback (most recent call last): 2025-12-04T09:53:29.8336653Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8337359Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8337856Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8338358Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8338869Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8339373Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8339717Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8339929Z 2025-12-04T09:53:29.8340169Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8341025Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8341671Z 2025-12-04T09:53:29.8341905Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8342409Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8344052Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8345604Z with policy(): 2025-12-04T09:53:29.8345914Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8347547Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8349081Z with policy(): 2025-12-04T09:53:29.8349500Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8349938Z Traceback (most recent call last): 2025-12-04T09:53:29.8350491Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8351102Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8351601Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8352107Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8352621Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8353130Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8353481Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8353762Z 2025-12-04T09:53:29.8353945Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8354772Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8355384Z 2025-12-04T09:53:29.8355605Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8356166Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8357815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8359388Z with policy(): 2025-12-04T09:53:29.8359699Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8361341Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8362886Z with policy(): 2025-12-04T09:53:29.8363262Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8363702Z Traceback (most recent call last): 2025-12-04T09:53:29.8364265Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8364899Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8365416Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8365934Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8367003Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8367562Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8367914Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8368128Z 2025-12-04T09:53:29.8368323Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8369109Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8369723Z 2025-12-04T09:53:29.8369946Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8370453Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8372169Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8373714Z with policy(): 2025-12-04T09:53:29.8374029Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8375673Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8377233Z with policy(): 2025-12-04T09:53:29.8377666Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8378103Z Traceback (most recent call last): 2025-12-04T09:53:29.8378668Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8379288Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8379835Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8380346Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8380873Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8381386Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8381731Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8381950Z 2025-12-04T09:53:29.8382131Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8383052Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8383663Z 2025-12-04T09:53:29.8383894Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8384399Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8386101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8387647Z with policy(): 2025-12-04T09:53:29.8387969Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8389628Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8391166Z with policy(): 2025-12-04T09:53:29.8391541Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8391984Z Traceback (most recent call last): 2025-12-04T09:53:29.8392545Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8393168Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8393674Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8394192Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8394712Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8395277Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8395633Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8395851Z 2025-12-04T09:53:29.8396044Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8396827Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8397447Z 2025-12-04T09:53:29.8397671Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8398181Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8399834Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8401467Z with policy(): 2025-12-04T09:53:29.8401785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8403470Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8405012Z with policy(): 2025-12-04T09:53:29.8405365Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8405847Z Traceback (most recent call last): 2025-12-04T09:53:29.8406414Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8407041Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8407616Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8408124Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8408646Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8409152Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8409506Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8409726Z 2025-12-04T09:53:29.8409907Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8410701Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8411312Z 2025-12-04T09:53:29.8411537Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8412051Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8413707Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8415300Z with policy(): 2025-12-04T09:53:29.8415616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8417309Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8418852Z with policy(): 2025-12-04T09:53:29.8419218Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8419660Z Traceback (most recent call last): 2025-12-04T09:53:29.8420221Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8420839Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8421343Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8421851Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8422371Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8422891Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8423294Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8423514Z 2025-12-04T09:53:29.8423700Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8424496Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8425097Z 2025-12-04T09:53:29.8425372Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8425882Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8427537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8429125Z with policy(): 2025-12-04T09:53:29.8429451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8431099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8432659Z with policy(): 2025-12-04T09:53:29.8433019Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8433458Z Traceback (most recent call last): 2025-12-04T09:53:29.8434022Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8434642Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8435155Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8435936Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8436453Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8436954Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8437304Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8437517Z 2025-12-04T09:53:29.8437705Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8438485Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8439092Z 2025-12-04T09:53:29.8439312Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8439823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8441572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8443118Z with policy(): 2025-12-04T09:53:29.8443428Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8445072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8446672Z with policy(): 2025-12-04T09:53:29.8447044Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8447576Z Traceback (most recent call last): 2025-12-04T09:53:29.8448201Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8448970Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8449532Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8450038Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8450557Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8451071Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8451416Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8451700Z 2025-12-04T09:53:29.8451885Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8452675Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8453281Z 2025-12-04T09:53:29.8453512Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8454021Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8455730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8457289Z with policy(): 2025-12-04T09:53:29.8457606Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8459260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8460798Z with policy(): 2025-12-04T09:53:29.8461167Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8461611Z Traceback (most recent call last): 2025-12-04T09:53:29.8462175Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8462801Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8463312Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8463832Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8464395Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8464914Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8465272Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8465491Z 2025-12-04T09:53:29.8465683Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8466470Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8467092Z 2025-12-04T09:53:29.8467318Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8467837Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8469510Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8471093Z with policy(): 2025-12-04T09:53:29.8471460Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8473110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8474706Z with policy(): 2025-12-04T09:53:29.8475066Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8475509Z Traceback (most recent call last): 2025-12-04T09:53:29.8476076Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8476699Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8477199Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8477721Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8478240Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8478747Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8479103Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8479328Z 2025-12-04T09:53:29.8479512Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8480308Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8480915Z 2025-12-04T09:53:29.8481141Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8481650Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8483303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8484903Z with policy(): 2025-12-04T09:53:29.8485219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8487023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8488613Z with policy(): 2025-12-04T09:53:29.8488986Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8489438Z Traceback (most recent call last): 2025-12-04T09:53:29.8489998Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8490626Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8491138Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8491645Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8492171Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8492735Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8493086Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8493303Z 2025-12-04T09:53:29.8493487Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8494323Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8494941Z 2025-12-04T09:53:29.8495166Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8495679Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8497320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8498930Z with policy(): 2025-12-04T09:53:29.8499255Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8500903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8502460Z with policy(): 2025-12-04T09:53:29.8502830Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8503281Z Traceback (most recent call last): 2025-12-04T09:53:29.8503836Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8504465Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8504971Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8505478Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8506008Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8506530Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8506900Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8514297Z 2025-12-04T09:53:29.8514511Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8515325Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8515954Z 2025-12-04T09:53:29.8516178Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8516708Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8518466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8520038Z with policy(): 2025-12-04T09:53:29.8520358Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8522011Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8523617Z with policy(): 2025-12-04T09:53:29.8523976Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8524411Z Traceback (most recent call last): 2025-12-04T09:53:29.8525070Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8525699Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8526205Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8526719Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8527314Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8527879Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8528243Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8528465Z 2025-12-04T09:53:29.8528648Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8529442Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8530049Z 2025-12-04T09:53:29.8530274Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8530783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8532446Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8534015Z with policy(): 2025-12-04T09:53:29.8534322Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8536391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8537992Z with policy(): 2025-12-04T09:53:29.8538356Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8538795Z Traceback (most recent call last): 2025-12-04T09:53:29.8539351Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8539976Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8540484Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8541123Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8541640Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8542143Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8542498Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8542710Z 2025-12-04T09:53:29.8542890Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8543673Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8544271Z 2025-12-04T09:53:29.8544501Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8545006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8546724Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8548313Z with policy(): 2025-12-04T09:53:29.8548625Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8550257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8551871Z with policy(): 2025-12-04T09:53:29.8552218Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8552657Z Traceback (most recent call last): 2025-12-04T09:53:29.8553210Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8553813Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8554312Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8554860Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8555394Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8555901Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8556244Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8556457Z 2025-12-04T09:53:29.8556643Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8557418Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8558024Z 2025-12-04T09:53:29.8558246Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8558748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8560391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8561943Z with policy(): 2025-12-04T09:53:29.8562255Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8563943Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8565527Z with policy(): 2025-12-04T09:53:29.8565884Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8566310Z Traceback (most recent call last): 2025-12-04T09:53:29.8566868Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8567530Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8568022Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8568580Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8569097Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8569613Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8569959Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8570177Z 2025-12-04T09:53:29.8570357Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8571187Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8571799Z 2025-12-04T09:53:29.8572025Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8572522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8574167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8575771Z with policy(): 2025-12-04T09:53:29.8576080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8577709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8579259Z with policy(): 2025-12-04T09:53:29.8579614Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8580047Z Traceback (most recent call last): 2025-12-04T09:53:29.8580596Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8581216Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8581721Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8582235Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8582740Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8583258Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8583602Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8583811Z 2025-12-04T09:53:29.8583988Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8584767Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8585382Z 2025-12-04T09:53:29.8585650Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8586151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8587792Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8589348Z with policy(): 2025-12-04T09:53:29.8589658Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8591290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8592867Z with policy(): 2025-12-04T09:53:29.8593212Z _______________ TestCudaMultiGPU.test_cuda_memory_leak_detection _______________ 2025-12-04T09:53:29.8593683Z Traceback (most recent call last): 2025-12-04T09:53:29.8594411Z File "/var/lib/jenkins/workspace/test/test_cuda_multigpu.py", line 1062, in test_cuda_memory_leak_detection 2025-12-04T09:53:29.8595280Z with self.assertRaisesRegex(RuntimeError, regex): 2025-12-04T09:53:29.8595777Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T09:53:29.8596281Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T09:53:29.8596869Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T09:53:29.8597373Z raise self.test_case.failureException(msg) 2025-12-04T09:53:29.8597721Z AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8597934Z 2025-12-04T09:53:29.8598120Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8598900Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8599501Z 2025-12-04T09:53:29.8599721Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8600231Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8601876Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8603436Z with policy(): 2025-12-04T09:53:29.8603746Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:53:29.8605437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py:3328: UserWarning: CUDA caching allocator reports a memory leak not verified by the driver API in __main__.TestCudaMultiGPU.test_cuda_memory_leak_detection! Caching allocator allocated memory was 0 and is now reported as 33554432 on device 0. CUDA driver allocated memory was 379781120 and is now 379781120. 2025-12-04T09:53:29.8606970Z with policy(): 2025-12-04T09:53:29.8607705Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_multigpu/test_cuda_multigpu-4895f15a6c4e5281.xml - 2025-12-04T09:53:29.8608474Z =========================== short test summary info ============================ 2025-12-04T09:53:29.8609217Z FAILED [0.1227s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8609749Z 2025-12-04T09:53:29.8609931Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8610716Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8611312Z 2025-12-04T09:53:29.8611541Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8612285Z FAILED [0.1022s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8612817Z 2025-12-04T09:53:29.8612998Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8613785Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8614435Z 2025-12-04T09:53:29.8614664Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8615408Z FAILED [0.1019s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8615936Z 2025-12-04T09:53:29.8616158Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8616937Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8617535Z 2025-12-04T09:53:29.8617760Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8618497Z FAILED [0.1023s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8619064Z 2025-12-04T09:53:29.8619247Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8620027Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8620623Z 2025-12-04T09:53:29.8620850Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8621594Z FAILED [0.1021s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8622121Z 2025-12-04T09:53:29.8622302Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8623079Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8623683Z 2025-12-04T09:53:29.8623907Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8624654Z FAILED [0.1003s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8625174Z 2025-12-04T09:53:29.8625355Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8626131Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8626727Z 2025-12-04T09:53:29.8626953Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8627694Z FAILED [0.0997s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8628216Z 2025-12-04T09:53:29.8628398Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8629178Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8629792Z 2025-12-04T09:53:29.8630014Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8630809Z FAILED [0.0998s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8631336Z 2025-12-04T09:53:29.8631516Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8632297Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8632903Z 2025-12-04T09:53:29.8633119Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8633870Z FAILED [0.0997s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8634395Z 2025-12-04T09:53:29.8634580Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8635703Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8636430Z 2025-12-04T09:53:29.8636657Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8637530Z FAILED [0.0997s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8638055Z 2025-12-04T09:53:29.8638238Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8639019Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8639621Z 2025-12-04T09:53:29.8639839Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8640645Z FAILED [0.0999s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8641165Z 2025-12-04T09:53:29.8641358Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8642125Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8642728Z 2025-12-04T09:53:29.8642950Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8643694Z FAILED [0.0998s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8644210Z 2025-12-04T09:53:29.8644392Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8645206Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8645815Z 2025-12-04T09:53:29.8646037Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8646785Z FAILED [0.0995s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8647382Z 2025-12-04T09:53:29.8647569Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8648337Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8648943Z 2025-12-04T09:53:29.8649159Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8649907Z FAILED [0.0999s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8650423Z 2025-12-04T09:53:29.8650611Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8651383Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8652057Z 2025-12-04T09:53:29.8652278Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8653023Z FAILED [0.0998s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8653543Z 2025-12-04T09:53:29.8653732Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8654499Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8655102Z 2025-12-04T09:53:29.8655320Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8656063Z FAILED [0.0997s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8656647Z 2025-12-04T09:53:29.8656833Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8657614Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8658214Z 2025-12-04T09:53:29.8658504Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8659254Z FAILED [0.0999s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8659770Z 2025-12-04T09:53:29.8659951Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8660725Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8661362Z 2025-12-04T09:53:29.8661585Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8662327Z FAILED [0.0996s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8662847Z 2025-12-04T09:53:29.8663025Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8663797Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8664386Z 2025-12-04T09:53:29.8664609Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8665401Z FAILED [0.0999s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8665922Z 2025-12-04T09:53:29.8666100Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8666870Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8667466Z 2025-12-04T09:53:29.8667690Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8668427Z FAILED [0.1000s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8668948Z 2025-12-04T09:53:29.8669131Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8669904Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8670510Z 2025-12-04T09:53:29.8670727Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8671469Z FAILED [0.0999s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8671989Z 2025-12-04T09:53:29.8672169Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8672987Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8673590Z 2025-12-04T09:53:29.8673806Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8674550Z FAILED [0.0997s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8675119Z 2025-12-04T09:53:29.8675297Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8676072Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8676672Z 2025-12-04T09:53:29.8676890Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8677678Z FAILED [0.0999s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8678202Z 2025-12-04T09:53:29.8678382Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8679207Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8679806Z 2025-12-04T09:53:29.8680027Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8680767Z FAILED [0.0998s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8681286Z 2025-12-04T09:53:29.8681480Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8682251Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8682904Z 2025-12-04T09:53:29.8683124Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8683873Z FAILED [0.0998s] test_cuda_multigpu.py::TestCudaMultiGPU::test_cuda_memory_leak_detection - AssertionError: RuntimeError not raised 2025-12-04T09:53:29.8684387Z 2025-12-04T09:53:29.8684574Z To execute this test, run the following from the base repo dir: 2025-12-04T09:53:29.8685341Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 PYTORCH_TEST_WITH_SLOW_GRADCHECK=1 python test/test_cuda_multigpu.py TestCudaMultiGPU.test_cuda_memory_leak_detection 2025-12-04T09:53:29.8685948Z 2025-12-04T09:53:29.8686168Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:53:29.8686632Z ======================== 25 failed, 25 passed in 7.73s ========================= 2025-12-04T09:53:29.8686884Z 2025-12-04T09:53:29.8687302Z FINISHED PRINTING LOG FILE of test_cuda_multigpu 1/1 (test/test-reports/test_cuda_multigpu_1.1_a23ea26f9721e68d_.log) 2025-12-04T09:53:29.8687768Z 2025-12-04T09:53:29.8688015Z Finished test_cuda_multigpu 1/1 ... [2025-12-04 09:53:29.817467][2512.233939853], took 0.19min 2025-12-04T09:53:29.8688924Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_multigpu/test_cuda_multigpu-4895f15a6c4e5281.xml 2025-12-04T09:53:30.5217383Z Uploading logs for 57118183167 to S3 2025-12-04T09:53:30.5785339Z Uploading artifacts took 0.67 seconds 2025-12-04T09:53:30.5785678Z test_cuda_multigpu 1/1 failed! 2025-12-04T09:53:30.5789920Z Running torch_np/numpy_tests/lib/test_index_tricks 1/1 ... [2025-12-04 09:53:30.578598][2512.99506852] 2025-12-04T09:53:30.5790671Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:30.5794480Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/lib/test_index_tricks.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:30.579038] 2025-12-04T09:53:33.8998700Z 2025-12-04T09:53:33.9000051Z torch_np/numpy_tests/lib/test_index_tricks 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.lib.test_index_tricks_1.1_529820d2d65bc59f_.log 2025-12-04T09:53:33.9000929Z Running 0 items in this shard: 2025-12-04T09:53:33.9001113Z 2025-12-04T09:53:33.9001473Z Finished torch_np/numpy_tests/lib/test_index_tricks 1/1 ... [2025-12-04 09:53:33.899546][2516.316014637], took 0.06min 2025-12-04T09:53:33.9131537Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_index_tricks/torch_np.numpy_tests.lib.test_index_tricks-506b926e701f49cf.xml 2025-12-04T09:53:33.9406109Z Running test_jit_autocast 1/1 ... [2025-12-04 09:53:33.940137][2516.356610432] 2025-12-04T09:53:33.9406712Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:33.9408832Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_jit_autocast.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:33.940460] 2025-12-04T09:53:39.4657614Z 2025-12-04T09:53:39.4658941Z test_jit_autocast 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_autocast_1.1_ec6ce096dc01ee34_.log 2025-12-04T09:53:39.4659699Z Running 0 items in this shard: 2025-12-04T09:53:39.4659939Z 2025-12-04T09:53:39.4660206Z Finished test_jit_autocast 1/1 ... [2025-12-04 09:53:39.465366][2521.881839934], took 0.09min 2025-12-04T09:53:39.4787861Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_jit_autocast/test_jit_autocast-94bb29b8a8a5a7c3.xml 2025-12-04T09:53:39.5040628Z Running test_xnnpack_integration 1/1 ... [2025-12-04 09:53:39.503745][2521.920218893] 2025-12-04T09:53:39.5041107Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:39.5044388Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_xnnpack_integration.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:39.504056] 2025-12-04T09:53:42.7746295Z 2025-12-04T09:53:42.7747255Z test_xnnpack_integration 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_xnnpack_integration_1.1_28da4c452e743da8_.log 2025-12-04T09:53:42.7747995Z Running 0 items in this shard: 2025-12-04T09:53:42.7748178Z 2025-12-04T09:53:42.7748460Z Finished test_xnnpack_integration 1/1 ... [2025-12-04 09:53:42.774283][2525.190756652], took 0.05min 2025-12-04T09:53:42.7880087Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_xnnpack_integration/test_xnnpack_integration-802360b65a8d44de.xml 2025-12-04T09:53:42.8184756Z Running nn/test_init 1/1 ... [2025-12-04 09:53:42.818152][2525.234625862] 2025-12-04T09:53:42.8185172Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:42.8189540Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_init.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:42.818459] 2025-12-04T09:53:46.4391904Z 2025-12-04T09:53:46.4392737Z nn/test_init 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_init_1.1_ec9b8b7537ee12ff_.log 2025-12-04T09:53:46.4393388Z Running 0 items in this shard: 2025-12-04T09:53:46.4393599Z 2025-12-04T09:53:46.4393823Z Finished nn/test_init 1/1 ... [2025-12-04 09:53:46.438879][2528.855352722], took 0.06min 2025-12-04T09:53:46.4526620Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_init/nn.test_init-da81cdd3c2c0ce98.xml 2025-12-04T09:53:46.4803131Z Running test_mobile_optimizer 1/1 ... [2025-12-04 09:53:46.479913][2528.896386891] 2025-12-04T09:53:46.4803776Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:46.4805761Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mobile_optimizer.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:46.480232] 2025-12-04T09:53:50.4013013Z 2025-12-04T09:53:50.4013805Z test_mobile_optimizer 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mobile_optimizer_1.1_cf6e1db294a4c552_.log 2025-12-04T09:53:50.4043862Z Running 100 items in this shard: test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures 2025-12-04T09:53:50.4072811Z 2025-12-04T09:53:50.4073084Z Finished test_mobile_optimizer 1/1 ... [2025-12-04 09:53:50.401082][2532.817555163], took 0.07min 2025-12-04T09:53:50.4153967Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-5cb9f1a76518723d.xml 2025-12-04T09:53:50.4597513Z Running test_type_promotion 1/1 ... [2025-12-04 09:53:50.459443][2532.875916825] 2025-12-04T09:53:50.4597962Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:50.4601353Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_type_promotion.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:50.459761] 2025-12-04T09:53:54.9317663Z 2025-12-04T09:53:54.9318518Z test_type_promotion 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_type_promotion_1.1_bfe6e4bccf03611c_.log 2025-12-04T09:53:54.9319228Z Running 0 items in this shard: 2025-12-04T09:53:54.9319422Z 2025-12-04T09:53:54.9319680Z Finished test_type_promotion 1/1 ... [2025-12-04 09:53:54.931411][2537.347882297], took 0.07min 2025-12-04T09:53:54.9459660Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_type_promotion/test_type_promotion-365dfbbc94f879c8.xml 2025-12-04T09:53:54.9708286Z Running test_reductions 1/1 ... [2025-12-04 09:53:54.970509][2537.38698211] 2025-12-04T09:53:54.9708699Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:53:54.9711722Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_reductions.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:53:54.970819] 2025-12-04T09:54:07.6070014Z 2025-12-04T09:54:07.6070853Z test_reductions 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_reductions_1.1_9d656aa654618631_.log 2025-12-04T09:54:07.6071529Z Running 0 items in this shard: 2025-12-04T09:54:07.6071752Z 2025-12-04T09:54:07.6071996Z Finished test_reductions 1/1 ... [2025-12-04 09:54:07.606662][2550.023135467], took 0.21min 2025-12-04T09:54:07.6208729Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_reductions/test_reductions-88d67cc570f30743.xml 2025-12-04T09:54:07.6845927Z Running test_autoload_disable 1/1 ... [2025-12-04 09:54:07.684227][2550.100699094] 2025-12-04T09:54:08.0120892Z Processing /var/lib/jenkins/workspace/test/cpp_extensions 2025-12-04T09:54:11.2208569Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T09:54:11.2228345Z [?25hBuilding wheels for collected packages: torch_test_cpp_extension 2025-12-04T09:55:36.8067990Z Building wheel for torch_test_cpp_extension (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - done 2025-12-04T09:55:36.8190086Z [?25h Created wheel for torch_test_cpp_extension: filename=torch_test_cpp_extension-0.0.0-cp310-cp310-linux_x86_64.whl size=13199558 sha256=3a42651fb92ae0c2b897b5aedf0a83a74f3f23436a1a39f5d6cdc4ea85d637af 2025-12-04T09:55:36.8193554Z Stored in directory: /tmp/pip-ephem-wheel-cache-nuxrpbkd/wheels/2b/79/8d/635cf291e138cfea331292ca746c62b61fade208eb55a7e3a1 2025-12-04T09:55:36.8209422Z Successfully built torch_test_cpp_extension 2025-12-04T09:55:37.1849162Z Installing collected packages: torch_test_cpp_extension 2025-12-04T09:55:37.4030922Z Successfully installed torch_test_cpp_extension-0.0.0 2025-12-04T09:55:39.9917751Z 2025-12-04T09:55:39.9918076Z Running tests... 2025-12-04T09:55:39.9918377Z ---------------------------------------------------------------------- 2025-12-04T09:55:40.3291416Z . 2025-12-04T09:55:40.3291707Z ---------------------------------------------------------------------- 2025-12-04T09:55:40.3292060Z Ran 1 test in 0.337s 2025-12-04T09:55:40.3292211Z 2025-12-04T09:55:40.3292295Z OK 2025-12-04T09:55:40.3292400Z 2025-12-04T09:55:40.3292501Z Generating XML reports... 2025-12-04T09:55:41.0396328Z Finished test_autoload_disable 1/1 ... [2025-12-04 09:55:41.039087][2643.455550572], took 1.56min 2025-12-04T09:55:41.0540423Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-unittest/test_autoload/TEST-TestDeviceBackendAutoload-20251204095539.xml 2025-12-04T09:55:44.8781023Z Running test batch 'tests to run' cost 1704.76 seconds 2025-12-04T09:55:44.8803860Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:44.8805319Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842144_66ce82b0d0f711f089f00242ac110002 2025-12-04T09:55:44.9748305Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842144_66ce82b0d0f711f089f00242ac110002 2025-12-04T09:55:44.9763214Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:44.9765234Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842144_66dd44eed0f711f089f00242ac110002 2025-12-04T09:55:45.0046872Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842144_66dd44eed0f711f089f00242ac110002 2025-12-04T09:55:45.0060267Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:45.0061680Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66e1cb36d0f711f089f00242ac110002 2025-12-04T09:55:45.0377875Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66e1cb36d0f711f089f00242ac110002 2025-12-04T09:55:45.0391353Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:45.0392578Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66e6d888d0f711f089f00242ac110002 2025-12-04T09:55:45.0718623Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66e6d888d0f711f089f00242ac110002 2025-12-04T09:55:45.0732963Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:45.0733932Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66ec0e84d0f711f089f00242ac110002 2025-12-04T09:55:45.1254713Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66ec0e84d0f711f089f00242ac110002 2025-12-04T09:55:45.1268360Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:45.1269638Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66f43aa0d0f711f089f00242ac110002 2025-12-04T09:55:45.1547522Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66f43aa0d0f711f089f00242ac110002 2025-12-04T09:55:45.1562213Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:45.1563587Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66f8b58ad0f711f089f00242ac110002 2025-12-04T09:55:45.1874292Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66f8b58ad0f711f089f00242ac110002 2025-12-04T09:55:45.1890266Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:45.1891393Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66fdb710d0f711f089f00242ac110002 2025-12-04T09:55:45.2216891Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_66fdb710d0f711f089f00242ac110002 2025-12-04T09:55:45.2234065Z Emitting td_test_failure_stats_v2 2025-12-04T09:55:45.2235166Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_6702f50ed0f711f089f00242ac110002 2025-12-04T09:55:45.2824349Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764842145_6702f50ed0f711f089f00242ac110002 2025-12-04T09:55:45.2825460Z inductor/test_triton_kernels 1/1 failed! 2025-12-04T09:55:45.2826096Z test_cuda 1/1 failed! 2025-12-04T09:55:45.2826464Z test_ci_sanity_check_fail 1/1 failed! 2025-12-04T09:55:45.2826908Z inductor/test_compiled_autograd 1/2 failed! 2025-12-04T09:55:45.2827356Z inductor/test_benchmark_fusion 1/1 failed! 2025-12-04T09:55:45.2827813Z inductor/test_cudacodecache 1/1 failed! 2025-12-04T09:55:45.2828264Z inductor/test_memory_planning 1/1 failed! 2025-12-04T09:55:45.2828692Z test_torch 1/1 failed! 2025-12-04T09:55:45.2829188Z test_cuda_multigpu 1/1 failed! 2025-12-04T09:55:46.0210365Z 2025-12-04T09:55:46.0210798Z real 28m31.158s 2025-12-04T09:55:46.0211056Z user 27m11.035s 2025-12-04T09:55:46.0211274Z sys 6m9.610s 2025-12-04T09:55:46.0211484Z + assert_git_not_dirty 2025-12-04T09:55:46.0211850Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *rocm* ]] 2025-12-04T09:55:46.0212338Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *xla* ]] 2025-12-04T09:55:46.0217848Z ++ git status --porcelain 2025-12-04T09:55:46.0218142Z ++ grep -v '?? third_party' 2025-12-04T09:55:50.2945244Z ++ true 2025-12-04T09:55:50.2947043Z + git_status= 2025-12-04T09:55:50.2947453Z + [[ -n '' ]] 2025-12-04T09:55:50.2947758Z + test_libtorch 2 2025-12-04T09:55:50.2948024Z + local SHARD=2 2025-12-04T09:55:50.2948259Z + [[ default != \s\l\o\w ]] 2025-12-04T09:55:50.2948526Z + echo 'Testing libtorch' 2025-12-04T09:55:50.2948781Z Testing libtorch 2025-12-04T09:55:50.2949506Z + ln -sf /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libbackend_with_compiler.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.2969342Z + ln -sf /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libjitbackend_test.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.2983405Z + ln -sf /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libcaffe2_nvrtc.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.2999824Z + ln -sf /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libc10.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libc10_cuda.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libc10d_cuda_test.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.3024357Z + ln -sf /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libshm /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libshm.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libshm_windows /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.3042293Z + ln -sf /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_cpu.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_cuda.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_cuda_linalg.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_global_deps.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_nvshmem.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_python.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorchbind_test.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.3058944Z + ln -sf '/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libnvfuser*' /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.3073080Z + export CPP_TESTS_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.3073734Z + CPP_TESTS_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:55:50.3074163Z + [[ -z 2 ]] 2025-12-04T09:55:50.3074370Z + [[ 2 == \1 ]] 2025-12-04T09:55:50.3074572Z + [[ -z 2 ]] 2025-12-04T09:55:50.3074774Z + [[ 2 == \2 ]] 2025-12-04T09:55:50.3074991Z + test_libtorch_jit 2025-12-04T09:55:50.3075215Z + pushd test 2025-12-04T09:55:50.3075450Z ~/workspace/test ~/workspace 2025-12-04T09:55:50.3075887Z + python cpp/jit/tests_setup.py setup 2025-12-04T09:55:52.2176484Z + popd 2025-12-04T09:55:52.2176730Z ~/workspace 2025-12-04T09:55:52.2177104Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *cuda* ]] 2025-12-04T09:55:52.2177515Z + [[ default != *nogpu* ]] 2025-12-04T09:55:52.2177772Z + LTC_TS_CUDA=1 2025-12-04T09:55:52.2178121Z + python test/run_test.py --cpp --verbose -i cpp/test_jit cpp/test_lazy 2025-12-04T09:55:57.1900281Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to /var/lib/jenkins/workspace/test/.pytorch-disabled-tests.json 2025-12-04T09:55:57.2004735Z Found test times from artifacts 2025-12-04T09:55:57.2402810Z Found test times from artifacts 2025-12-04T09:55:57.2414580Z Running all tests 2025-12-04T09:55:57.2418444Z Running parallel tests on 1 processes 2025-12-04T09:55:57.2418771Z Name: tests to run (est. time: 0.0min) 2025-12-04T09:55:57.2419216Z Serial tests (2): 2025-12-04T09:55:57.2419450Z cpp/test_jit 1/1 2025-12-04T09:55:57.2419687Z cpp/test_lazy 1/1 2025-12-04T09:55:57.2419928Z Parallel tests (0): 2025-12-04T09:55:57.2420186Z Name: excluded (est. time: 0.0min) 2025-12-04T09:55:57.2420468Z Serial tests (0): 2025-12-04T09:55:57.2420700Z Parallel tests (0): 2025-12-04T09:55:57.2425759Z Running cpp/test_jit 1/1 ... [2025-12-04 09:55:57.242290][2659.65876389] 2025-12-04T09:55:57.2426166Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:55:57.2428143Z Skipping C++ tests when running under RERUN_DISABLED_TESTS mode 2025-12-04T09:55:57.2428665Z Finished cpp/test_jit 1/1 ... [2025-12-04 09:55:57.242464][2659.658939723], took 0.00min 2025-12-04T09:55:57.9535398Z Uploading artifacts took 0.70 seconds 2025-12-04T09:55:57.9535865Z Running cpp/test_lazy 1/1 ... [2025-12-04 09:55:57.952988][2660.369460516] 2025-12-04T09:55:57.9536288Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:55:57.9536666Z Skipping C++ tests when running under RERUN_DISABLED_TESTS mode 2025-12-04T09:55:57.9537220Z Finished cpp/test_lazy 1/1 ... [2025-12-04 09:55:57.953170][2660.369645099], took 0.00min 2025-12-04T09:56:01.7366074Z Running test batch 'tests to run' cost 4.49 seconds 2025-12-04T09:56:02.4623296Z + pushd test 2025-12-04T09:56:02.4623615Z ~/workspace/test ~/workspace 2025-12-04T09:56:02.4623932Z + python cpp/jit/tests_setup.py shutdown 2025-12-04T09:56:04.1294430Z + popd 2025-12-04T09:56:04.1295271Z ~/workspace 2025-12-04T09:56:04.1295539Z + assert_git_not_dirty 2025-12-04T09:56:04.1295933Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *rocm* ]] 2025-12-04T09:56:04.1296432Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *xla* ]] 2025-12-04T09:56:04.1303211Z ++ git status --porcelain 2025-12-04T09:56:04.1304013Z ++ grep -v '?? third_party' 2025-12-04T09:56:04.5427920Z ++ true 2025-12-04T09:56:04.5429877Z + git_status= 2025-12-04T09:56:04.5430310Z + [[ -n '' ]] 2025-12-04T09:56:04.5430570Z + test_aot_compilation 2025-12-04T09:56:04.5430872Z + echo 'Testing Ahead of Time compilation' 2025-12-04T09:56:04.5431207Z Testing Ahead of Time compilation 2025-12-04T09:56:04.5433477Z + ln -sf /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libc10.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libc10_cuda.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libc10d_cuda_test.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:56:04.5455319Z + ln -sf /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_cpu.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_cuda.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_cuda_linalg.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_global_deps.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_nvshmem.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorch_python.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib/libtorchbind_test.so /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:56:04.5477563Z + '[' -f /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin/test_mobile_nnc ']' 2025-12-04T09:56:04.5478242Z + '[' -f /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin/aot_model_compiler_test ']' 2025-12-04T09:56:04.5478732Z + test_custom_script_ops 2025-12-04T09:56:04.5479018Z + echo 'Testing custom script operators' 2025-12-04T09:56:04.5479440Z Testing custom script operators 2025-12-04T09:56:04.5479829Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck == *s390x* ]] 2025-12-04T09:56:04.5480419Z + CUSTOM_OP_BUILD=/var/lib/jenkins/workspace/build/custom_test_artifacts/custom-op-build 2025-12-04T09:56:04.5480918Z + pushd test/custom_operator 2025-12-04T09:56:04.5481224Z ~/workspace/test/custom_operator ~/workspace 2025-12-04T09:56:04.5481715Z + cp -a /var/lib/jenkins/workspace/build/custom_test_artifacts/custom-op-build build 2025-12-04T09:56:04.5686735Z + python test_custom_ops.py -v 2025-12-04T09:56:07.2156960Z Test results will be stored in test-reports/python-unittest/test_custom_ops 2025-12-04T09:56:07.2171792Z 2025-12-04T09:56:07.2172183Z Running tests... 2025-12-04T09:56:07.2172501Z ---------------------------------------------------------------------- 2025-12-04T09:56:07.7985238Z test_abstract_impl_pystub_faketensor (__main__.TestCustomOperators) ... /var/lib/jenkins/workspace/test/custom_operator/my_custom_ops.py:13: FutureWarning: `create_unbacked_symint` is deprecated, please use `new_dynamic_size` instead 2025-12-04T09:56:07.7986982Z nnz = ctx.create_unbacked_symint() 2025-12-04T09:56:07.8070197Z ok (0.589s) 2025-12-04T09:56:07.9437181Z test_abstract_impl_pystub_meta (__main__.TestCustomOperators) ... ok (0.006s) 2025-12-04T09:56:07.9460119Z test_calling_custom_op (__main__.TestCustomOperators) ... ok (0.002s) 2025-12-04T09:56:07.9890226Z test_calling_custom_op_inside_script_module (__main__.TestCustomOperators) ... ok (0.043s) 2025-12-04T09:56:07.9901048Z test_calling_custom_op_string (__main__.TestCustomOperators) ... ok (0.001s) 2025-12-04T09:56:07.9928419Z test_calling_custom_op_with_autograd (__main__.TestCustomOperators) ... /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/autograd/graph.py:865: UserWarning: Using backward() with create_graph=True will create a reference cycle between the parameter and its gradient which can cause a memory leak. We recommend using autograd.grad when creating the graph to avoid this. If you have to use this function, make sure to reset the .grad fields of your parameters to None after use to break the cycle and avoid the leak. (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/autograd/engine.cpp:1301.) 2025-12-04T09:56:07.9930953Z return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass 2025-12-04T09:56:07.9942179Z ok (0.004s) 2025-12-04T09:56:07.9953162Z test_calling_custom_op_with_autograd_in_nograd_mode (__main__.TestCustomOperators) ... ok (0.001s) 2025-12-04T09:56:07.9959924Z test_custom_library_is_loaded (__main__.TestCustomOperators) ... ok (0.001s) 2025-12-04T09:56:08.0499382Z test_dynamo_pystub_suggestion (__main__.TestCustomOperators) ... ok (0.054s) 2025-12-04T09:56:08.0513095Z test_op_with_incorrect_abstract_impl_pystub (__main__.TestCustomOperators) ... ok (0.001s) 2025-12-04T09:56:08.0525606Z test_op_with_no_abstract_impl_pystub (__main__.TestCustomOperators) ... ok (0.001s) 2025-12-04T09:56:08.0609038Z test_saving_and_loading_script_module_with_custom_op (__main__.TestCustomOperators) ... ok (0.008s) 2025-12-04T09:56:08.0609475Z 2025-12-04T09:56:08.0609627Z ---------------------------------------------------------------------- 2025-12-04T09:56:08.0609981Z Ran 12 tests in 0.844s 2025-12-04T09:56:08.0610139Z 2025-12-04T09:56:08.0610226Z OK 2025-12-04T09:56:08.0610336Z 2025-12-04T09:56:08.0610440Z Generating XML reports... 2025-12-04T09:56:08.0647964Z Generated XML report: test-reports/python-unittest/test_custom_ops/TEST-TestCustomOperators-20251204095607.xml 2025-12-04T09:56:08.8798751Z + python model.py --export-script-module=model.pt 2025-12-04T09:56:10.5561889Z + build/test_custom_ops ./model.pt 2025-12-04T09:56:10.9081458Z [W1204 09:56:10.324005998 engine.cpp:1301] Warning: Using backward() with create_graph=True will create a reference cycle between the parameter and its gradient which can cause a memory leak. We recommend using autograd.grad when creating the graph to avoid this. If you have to use this function, make sure to reset the .grad fields of your parameters to None after use to break the cycle and avoid the leak. (function operator()) 2025-12-04T09:56:11.1944057Z ok 2025-12-04T09:56:11.4214913Z + popd 2025-12-04T09:56:11.4215157Z ~/workspace 2025-12-04T09:56:11.4215369Z + assert_git_not_dirty 2025-12-04T09:56:11.4215751Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *rocm* ]] 2025-12-04T09:56:11.4216244Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *xla* ]] 2025-12-04T09:56:11.4225520Z ++ git status --porcelain 2025-12-04T09:56:11.4225924Z ++ grep -v '?? third_party' 2025-12-04T09:56:11.8358859Z ++ true 2025-12-04T09:56:11.8359118Z + git_status= 2025-12-04T09:56:11.8359365Z + [[ -n '' ]] 2025-12-04T09:56:11.8359573Z + test_custom_backend 2025-12-04T09:56:11.8359827Z + echo 'Testing custom backends' 2025-12-04T09:56:11.8360110Z Testing custom backends 2025-12-04T09:56:11.8361069Z + CUSTOM_BACKEND_BUILD=/var/lib/jenkins/workspace/build/custom_test_artifacts/custom-backend-build 2025-12-04T09:56:11.8361818Z + pushd test/custom_backend 2025-12-04T09:56:11.8362204Z ~/workspace/test/custom_backend ~/workspace 2025-12-04T09:56:11.8362826Z + cp -a /var/lib/jenkins/workspace/build/custom_test_artifacts/custom-backend-build build 2025-12-04T09:56:11.8564127Z + python test_custom_backend.py -v 2025-12-04T09:56:14.4825016Z Test results will be stored in test-reports/python-unittest/test_custom_backend 2025-12-04T09:56:14.4836422Z 2025-12-04T09:56:14.4838553Z Running tests... 2025-12-04T09:56:14.4838868Z ---------------------------------------------------------------------- 2025-12-04T09:56:14.4843590Z test_execute (__main__.TestCustomBackend) 2025-12-04T09:56:14.7357998Z Test execution using the custom backend. ... ok (0.252s) 2025-12-04T09:56:14.8589609Z test_save_load (__main__.TestCustomBackend) 2025-12-04T09:56:14.8792502Z Test that a lowered module can be executed correctly ... ok (0.021s) 2025-12-04T09:56:14.8793136Z 2025-12-04T09:56:14.8793456Z ---------------------------------------------------------------------- 2025-12-04T09:56:14.8794141Z Ran 2 tests in 0.396s 2025-12-04T09:56:14.8794454Z 2025-12-04T09:56:14.8794605Z OK 2025-12-04T09:56:14.8794823Z 2025-12-04T09:56:14.8795027Z Generating XML reports... 2025-12-04T09:56:14.8820085Z Generated XML report: test-reports/python-unittest/test_custom_backend/TEST-TestCustomBackend-20251204095614.xml 2025-12-04T09:56:15.6050065Z + python backend.py --export-module-to=model.pt 2025-12-04T09:56:17.3259250Z + build/test_custom_backend ./model.pt 2025-12-04T09:56:17.6641014Z Testing custom_backend 2025-12-04T09:56:17.7216851Z OK 2025-12-04T09:56:17.8673911Z + rm -f ./model.pt 2025-12-04T09:56:17.8706497Z + popd 2025-12-04T09:56:17.8707047Z ~/workspace 2025-12-04T09:56:17.8707294Z + assert_git_not_dirty 2025-12-04T09:56:17.8707659Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *rocm* ]] 2025-12-04T09:56:17.8708139Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *xla* ]] 2025-12-04T09:56:17.8715120Z ++ git status --porcelain 2025-12-04T09:56:17.8715591Z ++ grep -v '?? third_party' 2025-12-04T09:56:18.2821010Z ++ true 2025-12-04T09:56:18.2822720Z + git_status= 2025-12-04T09:56:18.2823146Z + [[ -n '' ]] 2025-12-04T09:56:18.2823487Z + test_torch_function_benchmark 2025-12-04T09:56:18.2823832Z + echo 'Testing __torch_function__ benchmarks' 2025-12-04T09:56:18.2824171Z Testing __torch_function__ benchmarks 2025-12-04T09:56:18.2824487Z + pushd benchmarks/overrides_benchmark 2025-12-04T09:56:18.2824901Z ~/workspace/benchmarks/overrides_benchmark ~/workspace 2025-12-04T09:56:18.2825496Z + python bench.py -n 1 -m 2 2025-12-04T09:56:19.6203371Z Type tensor had a minimum time of 0.0059604644775390625 us and a standard deviation of 0.37038649315945804 us. 2025-12-04T09:56:19.6204158Z Type SubTensor had a minimum time of 0.013589859008789062 us and a standard deviation of 0.04214685031911358 us. 2025-12-04T09:56:19.6205264Z Type WithTorchFunction had a minimum time of 0.0064373016357421875 us and a standard deviation of 0.015341453035944141 us. 2025-12-04T09:56:19.6206119Z Type SubWithTorchFunction had a minimum time of 0.010967254638671875 us and a standard deviation of 0.006406321062968345 us. 2025-12-04T09:56:19.9521925Z + python pyspybench.py Tensor -n 1 2025-12-04T09:56:21.6212720Z + python pyspybench.py SubTensor -n 1 2025-12-04T09:56:23.2893501Z + python pyspybench.py WithTorchFunction -n 1 2025-12-04T09:56:24.9478663Z + python pyspybench.py SubWithTorchFunction -n 1 2025-12-04T09:56:26.6084777Z + popd 2025-12-04T09:56:26.6085031Z ~/workspace 2025-12-04T09:56:26.6085276Z + assert_git_not_dirty 2025-12-04T09:56:26.6085644Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *rocm* ]] 2025-12-04T09:56:26.6086146Z + [[ linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck != *xla* ]] 2025-12-04T09:56:26.6093328Z ++ git status --porcelain 2025-12-04T09:56:26.6093641Z ++ grep -v '?? third_party' 2025-12-04T09:56:27.0217067Z ++ true 2025-12-04T09:56:27.0218864Z + git_status= 2025-12-04T09:56:27.0219264Z + [[ -n '' ]] 2025-12-04T09:56:27.0220408Z + sccache_epilogue 2025-12-04T09:56:27.0220780Z + echo '::group::Sccache Compilation Log' 2025-12-04T09:56:27.0221688Z ##[group]Sccache Compilation Log 2025-12-04T09:56:27.0222056Z + echo '=================== sccache compilation log ===================' 2025-12-04T09:56:27.0222452Z =================== sccache compilation log =================== 2025-12-04T09:56:27.0223054Z + python /var/lib/jenkins/workspace/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-12-04T09:56:27.0374613Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-12-04T09:56:27.0375310Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-12-04T09:56:27.0375773Z + sccache --show-stats 2025-12-04T09:56:27.0412495Z Compile requests 716 2025-12-04T09:56:27.0412967Z Compile requests executed 265 2025-12-04T09:56:27.0413380Z Cache hits 116 2025-12-04T09:56:27.0413772Z Cache hits (C/C++) 116 2025-12-04T09:56:27.0414165Z Cache misses 149 2025-12-04T09:56:27.0414550Z Cache misses (C/C++) 149 2025-12-04T09:56:27.0414963Z Cache hits rate 43.77 % 2025-12-04T09:56:27.0415392Z Cache hits rate (C/C++) 43.77 % 2025-12-04T09:56:27.0415804Z Cache timeouts 0 2025-12-04T09:56:27.0416199Z Cache read errors 0 2025-12-04T09:56:27.0416583Z Forced recaches 0 2025-12-04T09:56:27.0416871Z Cache write errors 0 2025-12-04T09:56:27.0417160Z Cache errors 0 2025-12-04T09:56:27.0417669Z Compilations 149 2025-12-04T09:56:27.0418029Z Compilation failures 0 2025-12-04T09:56:27.0418402Z Non-cacheable compilations 0 2025-12-04T09:56:27.0418712Z Non-cacheable calls 103 2025-12-04T09:56:27.0419094Z Non-compilation calls 348 2025-12-04T09:56:27.0419400Z Unsupported compiler calls 0 2025-12-04T09:56:27.0419717Z Average cache write 0.044 s 2025-12-04T09:56:27.0420029Z Average compiler 6.237 s 2025-12-04T09:56:27.0420331Z Average cache read hit 0.027 s 2025-12-04T09:56:27.0420650Z Failed distributed compilations 0 2025-12-04T09:56:27.0420859Z 2025-12-04T09:56:27.0420959Z Non-cacheable reasons: 2025-12-04T09:56:27.0421221Z unknown source language 100 2025-12-04T09:56:27.0421622Z -E 3 2025-12-04T09:56:27.0421820Z 2025-12-04T09:56:27.0422049Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T09:56:27.0422485Z Version (client) 0.10.0 2025-12-04T09:56:27.0422776Z + sccache --stop-server 2025-12-04T09:56:27.0443955Z Stopping sccache server... 2025-12-04T09:56:27.0448196Z Compile requests 716 2025-12-04T09:56:27.0448766Z Compile requests executed 265 2025-12-04T09:56:27.0449077Z Cache hits 116 2025-12-04T09:56:27.0449436Z Cache hits (C/C++) 116 2025-12-04T09:56:27.0449732Z Cache misses 149 2025-12-04T09:56:27.0450011Z Cache misses (C/C++) 149 2025-12-04T09:56:27.0450382Z Cache hits rate 43.77 % 2025-12-04T09:56:27.0450690Z Cache hits rate (C/C++) 43.77 % 2025-12-04T09:56:27.0451244Z Cache timeouts 0 2025-12-04T09:56:27.0451538Z Cache read errors 0 2025-12-04T09:56:27.0451857Z Forced recaches 0 2025-12-04T09:56:27.0452210Z Cache write errors 0 2025-12-04T09:56:27.0452635Z Cache errors 0 2025-12-04T09:56:27.0453040Z Compilations 149 2025-12-04T09:56:27.0453449Z Compilation failures 0 2025-12-04T09:56:27.0453851Z Non-cacheable compilations 0 2025-12-04T09:56:27.0454284Z Non-cacheable calls 103 2025-12-04T09:56:27.0454650Z Non-compilation calls 348 2025-12-04T09:56:27.0454961Z Unsupported compiler calls 0 2025-12-04T09:56:27.0455330Z Average cache write 0.044 s 2025-12-04T09:56:27.0455641Z Average compiler 6.237 s 2025-12-04T09:56:27.0455945Z Average cache read hit 0.027 s 2025-12-04T09:56:27.0456343Z Failed distributed compilations 0 2025-12-04T09:56:27.0456559Z 2025-12-04T09:56:27.0456660Z Non-cacheable reasons: 2025-12-04T09:56:27.0456926Z unknown source language 100 2025-12-04T09:56:27.0457211Z -E 3 2025-12-04T09:56:27.0457402Z 2025-12-04T09:56:27.0457700Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T09:56:27.0458143Z Version (client) 0.10.0 2025-12-04T09:56:27.0458453Z + echo ::endgroup:: 2025-12-04T09:56:27.0458942Z ##[endgroup] 2025-12-04T09:56:27.0459169Z + cleanup_workspace 2025-12-04T09:56:27.0459646Z + echo 'sudo may print the following warning message that can be ignored. The chown command will still run.' 2025-12-04T09:56:27.0460467Z sudo may print the following warning message that can be ignored. The chown command will still run. 2025-12-04T09:56:27.0461065Z + echo ' sudo: setrlimit(RLIMIT_STACK): Operation not permitted' 2025-12-04T09:56:27.0461520Z sudo: setrlimit(RLIMIT_STACK): Operation not permitted 2025-12-04T09:56:27.0462051Z + echo 'For more details refer to https://github.com/sudo-project/sudo/issues/42' 2025-12-04T09:56:27.0462621Z For more details refer to https://github.com/sudo-project/sudo/issues/42 2025-12-04T09:56:27.0463203Z + sudo chown -R 1000 /var/lib/jenkins/workspace 2025-12-04T09:56:28.1186729Z ##[group]Run pytorch/test-infra/.github/actions/upload-benchmark-results@main 2025-12-04T09:56:28.1187189Z with: 2025-12-04T09:56:28.1187435Z benchmark-results-dir: test/test-reports 2025-12-04T09:56:28.1187751Z dry-run: false 2025-12-04T09:56:28.1187975Z schema-version: v3 2025-12-04T09:56:28.1188402Z github-token: *** 2025-12-04T09:56:28.1188621Z env: 2025-12-04T09:56:28.1188823Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:56:28.1189083Z HAS_NVIDIA_GPU: true 2025-12-04T09:56:28.1189391Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:56:28.1189909Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:56:28.1190375Z ##[endgroup] 2025-12-04T09:56:28.1208245Z ##[group]Run set -eux 2025-12-04T09:56:28.1208627Z set -eux 2025-12-04T09:56:28.1208843Z  2025-12-04T09:56:28.1209045Z if [[ -n "" ]]; then 2025-12-04T09:56:28.1209306Z  source "" 2025-12-04T09:56:28.1209553Z fi 2025-12-04T09:56:28.1209883Z python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-12-04T09:56:28.1210283Z  2025-12-04T09:56:28.1210490Z DEVICE_NAME="" 2025-12-04T09:56:28.1210739Z DEVICE_TYPE="" 2025-12-04T09:56:28.1210977Z  2025-12-04T09:56:28.1211204Z if command -v nvidia-smi; then 2025-12-04T09:56:28.1211643Z  # NB: I'm using PyTorch here to get the device name, however, it needs to 2025-12-04T09:56:28.1212223Z  # install the correct version of PyTorch manually for now. Any PyTorch 2025-12-04T09:56:28.1212774Z  # version is fine, I just use 2.7.1 to satify PYPIDEP linter 2025-12-04T09:56:28.1213289Z  python3 -mpip install torch==2.7.1 2025-12-04T09:56:28.1213633Z elif command -v rocminfo; then 2025-12-04T09:56:28.1214066Z  # NB: Installing torch on ROCm runner with pip here causes CI to fail 2025-12-04T09:56:28.1214616Z  # with a memoryview is too large error only on MI300 runners. Is pip 2025-12-04T09:56:28.1215164Z  # version on ROCm runner there too old? As a workaround, let's use the 2025-12-04T09:56:28.1215639Z  # GPU device name coming from rocminfo instead 2025-12-04T09:56:28.1215988Z  DEVICE_NAME=rocm 2025-12-04T09:56:28.1216457Z  DEVICE_TYPE=$(rocminfo | grep "Marketing Name" | tail -n1 | awk -F':' '{print $2}' | xargs) 2025-12-04T09:56:28.1216926Z fi 2025-12-04T09:56:28.1217137Z  2025-12-04T09:56:28.1217402Z echo "DEVICE_NAME=$DEVICE_NAME" >> $GITHUB_ENV 2025-12-04T09:56:28.1217802Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-12-04T09:56:28.1232363Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:56:28.1232720Z env: 2025-12-04T09:56:28.1232925Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:56:28.1233173Z HAS_NVIDIA_GPU: true 2025-12-04T09:56:28.1233486Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:56:28.1234011Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:56:28.1234468Z ##[endgroup] 2025-12-04T09:56:28.1276740Z + [[ -n '' ]] 2025-12-04T09:56:28.1277086Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-12-04T09:56:28.3831628Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:56:29.6578533Z Collecting boto3==1.35.33 2025-12-04T09:56:29.6737393Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-12-04T09:56:30.0450508Z Collecting psutil==7.0.0 2025-12-04T09:56:30.0488656Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-12-04T09:56:30.0797598Z Collecting pynvml==12.0.0 2025-12-04T09:56:30.0834834Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-12-04T09:56:30.1357225Z Collecting s3transfer<0.11.0,>=0.10.0 2025-12-04T09:56:30.1388611Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-12-04T09:56:31.4904420Z Collecting botocore<1.36.0,>=1.35.33 2025-12-04T09:56:31.4940211Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-12-04T09:56:31.6330574Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-12-04T09:56:31.6784939Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-12-04T09:56:31.6816367Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-12-04T09:56:31.6903566Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-12-04T09:56:31.6910986Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-12-04T09:56:31.8512254Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-12-04T09:56:31.9892260Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-12-04T09:56:32.5645019Z Attempting uninstall: nvidia-ml-py 2025-12-04T09:56:32.5646037Z Found existing installation: nvidia-ml-py 11.525.84 2025-12-04T09:56:32.5661462Z Uninstalling nvidia-ml-py-11.525.84: 2025-12-04T09:56:32.5918540Z Successfully uninstalled nvidia-ml-py-11.525.84 2025-12-04T09:56:32.6529503Z Attempting uninstall: psutil 2025-12-04T09:56:32.6530096Z Found existing installation: psutil 5.9.8 2025-12-04T09:56:32.6620118Z Uninstalling psutil-5.9.8: 2025-12-04T09:56:32.6627648Z Successfully uninstalled psutil-5.9.8 2025-12-04T09:56:32.8343711Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-12-04T09:56:32.9667664Z + DEVICE_NAME= 2025-12-04T09:56:32.9667903Z + DEVICE_TYPE= 2025-12-04T09:56:32.9668362Z /usr/bin/nvidia-smi 2025-12-04T09:56:32.9668764Z + command -v nvidia-smi 2025-12-04T09:56:32.9669047Z + python3 -mpip install torch==2.7.1 2025-12-04T09:56:33.2056812Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:56:33.5067302Z Collecting torch==2.7.1 2025-12-04T09:56:33.5253739Z Downloading torch-2.7.1-cp39-cp39-manylinux_2_28_x86_64.whl (821.1 MB) 2025-12-04T09:56:45.4515475Z Collecting filelock 2025-12-04T09:56:45.4547248Z Downloading filelock-3.19.1-py3-none-any.whl (15 kB) 2025-12-04T09:56:45.4762288Z Collecting nvidia-cufile-cu12==1.11.1.6 2025-12-04T09:56:45.4797184Z Downloading nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.1 MB) 2025-12-04T09:56:45.5628050Z Collecting networkx 2025-12-04T09:56:45.5663228Z Downloading networkx-3.2.1-py3-none-any.whl (1.6 MB) 2025-12-04T09:56:45.6223777Z Collecting nvidia-nvtx-cu12==12.6.77 2025-12-04T09:56:45.6260231Z Downloading nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (89 kB) 2025-12-04T09:56:45.6801022Z Collecting triton==3.3.1 2025-12-04T09:56:45.6864184Z Downloading triton-3.3.1-cp39-cp39-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (155.6 MB) 2025-12-04T09:56:47.2375549Z Collecting nvidia-cuda-nvrtc-cu12==12.6.77 2025-12-04T09:56:47.2408340Z Downloading nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl (23.7 MB) 2025-12-04T09:56:47.4605549Z Requirement already satisfied: typing-extensions>=4.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from torch==2.7.1) (4.15.0) 2025-12-04T09:56:47.4911291Z Collecting nvidia-cuda-runtime-cu12==12.6.77 2025-12-04T09:56:47.4970577Z Downloading nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (897 kB) 2025-12-04T09:56:47.5513545Z Collecting nvidia-cudnn-cu12==9.5.1.17 2025-12-04T09:56:47.5547503Z Downloading nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_x86_64.whl (571.0 MB) 2025-12-04T09:56:55.5512250Z Collecting nvidia-nvjitlink-cu12==12.6.85 2025-12-04T09:56:55.5571343Z Downloading nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl (19.7 MB) 2025-12-04T09:56:55.7451274Z Collecting nvidia-cusparselt-cu12==0.6.3 2025-12-04T09:56:55.7483474Z Downloading nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl (156.8 MB) 2025-12-04T09:56:57.2826306Z Collecting nvidia-curand-cu12==10.3.7.77 2025-12-04T09:56:57.2857560Z Downloading nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (56.3 MB) 2025-12-04T09:56:57.8479110Z Collecting nvidia-cufft-cu12==11.3.0.4 2025-12-04T09:56:57.8513804Z Downloading nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (200.2 MB) 2025-12-04T09:57:00.0117684Z Collecting sympy>=1.13.3 2025-12-04T09:57:00.0152565Z Downloading sympy-1.14.0-py3-none-any.whl (6.3 MB) 2025-12-04T09:57:00.1068544Z Collecting nvidia-cuda-cupti-cu12==12.6.80 2025-12-04T09:57:00.1283167Z Downloading nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (8.9 MB) 2025-12-04T09:57:00.2313207Z Collecting nvidia-cublas-cu12==12.6.4.1 2025-12-04T09:57:00.2369945Z Downloading nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (393.1 MB) 2025-12-04T09:57:05.3770555Z Collecting nvidia-cusparse-cu12==12.5.4.2 2025-12-04T09:57:05.3826178Z Downloading nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (216.6 MB) 2025-12-04T09:57:07.7897072Z WARNING: Retrying (Retry(total=4, connect=None, read=None, redirect=None, status=None)) after connection broken by 'ProtocolError('Connection aborted.', ConnectionResetError(104, 'Connection reset by peer'))': /simple/nvidia-cusolver-cu12/ 2025-12-04T09:57:07.8336717Z Collecting nvidia-cusolver-cu12==11.7.1.2 2025-12-04T09:57:07.8371277Z Downloading nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (158.2 MB) 2025-12-04T09:57:09.4106774Z Collecting fsspec 2025-12-04T09:57:09.4146968Z Downloading fsspec-2025.10.0-py3-none-any.whl (200 kB) 2025-12-04T09:57:09.4216520Z Requirement already satisfied: jinja2 in /usr/lib/python3.9/site-packages (from torch==2.7.1) (2.11.3) 2025-12-04T09:57:09.4488798Z Collecting nvidia-nccl-cu12==2.26.2 2025-12-04T09:57:09.4522205Z Downloading nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (201.3 MB) 2025-12-04T09:57:11.6432179Z Requirement already satisfied: setuptools>=40.8.0 in /usr/lib/python3.9/site-packages (from triton==3.3.1->torch==2.7.1) (59.6.0) 2025-12-04T09:57:11.6753457Z Collecting mpmath<1.4,>=1.1.0 2025-12-04T09:57:11.6786655Z Downloading mpmath-1.3.0-py3-none-any.whl (536 kB) 2025-12-04T09:57:11.7735897Z Requirement already satisfied: MarkupSafe>=0.23 in /usr/lib64/python3.9/site-packages (from jinja2->torch==2.7.1) (1.1.1) 2025-12-04T09:57:12.1319088Z Installing collected packages: nvidia-nvjitlink-cu12, nvidia-cusparse-cu12, nvidia-cublas-cu12, mpmath, triton, sympy, nvidia-nvtx-cu12, nvidia-nccl-cu12, nvidia-cusparselt-cu12, nvidia-cusolver-cu12, nvidia-curand-cu12, nvidia-cufile-cu12, nvidia-cufft-cu12, nvidia-cudnn-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, networkx, fsspec, filelock, torch 2025-12-04T09:57:21.4869369Z WARNING: The scripts proton and proton-viewer are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-12-04T09:57:21.4870216Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-12-04T09:57:25.5485986Z WARNING: The script isympy is installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-12-04T09:57:25.5487249Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-12-04T09:57:56.9423717Z WARNING: The scripts torchfrtrace and torchrun are installed in '/home/ec2-user/.local/bin' which is not on PATH. 2025-12-04T09:57:56.9424905Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-12-04T09:57:57.1488780Z Successfully installed filelock-3.19.1 fsspec-2025.10.0 mpmath-1.3.0 networkx-3.2.1 nvidia-cublas-cu12-12.6.4.1 nvidia-cuda-cupti-cu12-12.6.80 nvidia-cuda-nvrtc-cu12-12.6.77 nvidia-cuda-runtime-cu12-12.6.77 nvidia-cudnn-cu12-9.5.1.17 nvidia-cufft-cu12-11.3.0.4 nvidia-cufile-cu12-1.11.1.6 nvidia-curand-cu12-10.3.7.77 nvidia-cusolver-cu12-11.7.1.2 nvidia-cusparse-cu12-12.5.4.2 nvidia-cusparselt-cu12-0.6.3 nvidia-nccl-cu12-2.26.2 nvidia-nvjitlink-cu12-12.6.85 nvidia-nvtx-cu12-12.6.77 sympy-1.14.0 torch-2.7.1 triton-3.3.1 2025-12-04T09:57:57.8329709Z + echo DEVICE_NAME= 2025-12-04T09:57:57.8330466Z + echo DEVICE_TYPE= 2025-12-04T09:57:57.8361238Z ##[group]Run set -eux 2025-12-04T09:57:57.8361506Z set -eux 2025-12-04T09:57:57.8361846Z  2025-12-04T09:57:57.8362077Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-12-04T09:57:57.8362429Z  echo "Missing github-token input" 2025-12-04T09:57:57.8362755Z  exit 1 2025-12-04T09:57:57.8362975Z fi 2025-12-04T09:57:57.8372586Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:57:57.8372952Z env: 2025-12-04T09:57:57.8373169Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:57:57.8373427Z HAS_NVIDIA_GPU: true 2025-12-04T09:57:57.8373743Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:57:57.8374281Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:57:57.8374786Z DEVICE_NAME: 2025-12-04T09:57:57.8375031Z DEVICE_TYPE: 2025-12-04T09:57:57.8375450Z GITHUB_TOKEN: *** 2025-12-04T09:57:57.8375677Z ##[endgroup] 2025-12-04T09:57:57.8409911Z + [[ -z *** ]] 2025-12-04T09:57:57.8486474Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-12-04T09:57:57.8486897Z with: 2025-12-04T09:57:57.8487489Z github-token: *** 2025-12-04T09:57:57.8487718Z env: 2025-12-04T09:57:57.8487935Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:57:57.8488214Z HAS_NVIDIA_GPU: true 2025-12-04T09:57:57.8488526Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:57:57.8489062Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:57:57.8489531Z DEVICE_NAME: 2025-12-04T09:57:57.8489753Z DEVICE_TYPE: 2025-12-04T09:57:57.8489968Z ##[endgroup] 2025-12-04T09:57:57.8523340Z ##[group]Run set -eux 2025-12-04T09:57:57.8523602Z set -eux 2025-12-04T09:57:57.8523816Z  2025-12-04T09:57:57.8524276Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T09:57:57.8533577Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:57:57.8533966Z env: 2025-12-04T09:57:57.8534184Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:57:57.8534442Z HAS_NVIDIA_GPU: true 2025-12-04T09:57:57.8534750Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:57:57.8535476Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:57:57.8535942Z DEVICE_NAME: 2025-12-04T09:57:57.8536153Z DEVICE_TYPE: 2025-12-04T09:57:57.8536496Z GITHUB_TOKEN: *** 2025-12-04T09:57:57.8536726Z ##[endgroup] 2025-12-04T09:57:57.8570389Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 19922826259 i-016559b86632fc7da 2025-12-04T09:58:00.3194273Z setting job-id=57118183167 2025-12-04T09:58:00.3195620Z setting job-name=linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:58:00.3320319Z ##[group]Run set -eux 2025-12-04T09:58:00.3320610Z set -eux 2025-12-04T09:58:00.3320827Z  2025-12-04T09:58:00.3321035Z if [[ -n "" ]]; then 2025-12-04T09:58:00.3321297Z  source "" 2025-12-04T09:58:00.3321517Z fi 2025-12-04T09:58:00.3321925Z  2025-12-04T09:58:00.3322301Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-12-04T09:58:00.3322800Z  --schema-version "${SCHEMA_VERSION}" \ 2025-12-04T09:58:00.3323135Z  --repo "${REPO}" \ 2025-12-04T09:58:00.3323429Z  --head-branch "${HEAD_BRANCH}" \ 2025-12-04T09:58:00.3323751Z  --head-sha "${HEAD_SHA}" \ 2025-12-04T09:58:00.3324071Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-12-04T09:58:00.3324417Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-12-04T09:58:00.3324738Z  --job-id "${JOB_ID}" \ 2025-12-04T09:58:00.3325028Z  --job-name "${JOB_NAME}" 2025-12-04T09:58:00.3335095Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:00.3335916Z env: 2025-12-04T09:58:00.3336126Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:00.3336379Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:00.3336694Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:00.3337238Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:00.3337700Z DEVICE_NAME: 2025-12-04T09:58:00.3337923Z DEVICE_TYPE: 2025-12-04T09:58:00.3338148Z SCHEMA_VERSION: v3 2025-12-04T09:58:00.3338383Z REPO: pytorch/pytorch 2025-12-04T09:58:00.3338643Z HEAD_BRANCH: refs/heads/main 2025-12-04T09:58:00.3338965Z HEAD_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:58:00.3339307Z WORKFLOW_RUN_ID: 19922826259 2025-12-04T09:58:00.3339563Z RUN_ATTEMPT: 1 2025-12-04T09:58:00.3339780Z JOB_ID: 57118183167 2025-12-04T09:58:00.3340440Z JOB_NAME: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:58:00.3341319Z ##[endgroup] 2025-12-04T09:58:00.3372792Z + [[ -n '' ]] 2025-12-04T09:58:00.3374921Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 --workflow-id 19922826259 --run-attempt 1 --job-id 57118183167 --job-name 'linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests)' 2025-12-04T09:58:00.3770961Z ##[group]Run set -eux 2025-12-04T09:58:00.3771225Z set -eux 2025-12-04T09:58:00.3771443Z  2025-12-04T09:58:00.3771646Z if [[ -n "" ]]; then 2025-12-04T09:58:00.3771904Z  source "" 2025-12-04T09:58:00.3772133Z fi 2025-12-04T09:58:00.3772332Z  2025-12-04T09:58:00.3772720Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-12-04T09:58:00.3781801Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:00.3782168Z env: 2025-12-04T09:58:00.3782399Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:00.3782668Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:00.3782981Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:00.3783512Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:00.3783988Z DEVICE_NAME: 2025-12-04T09:58:00.3784210Z DEVICE_TYPE: 2025-12-04T09:58:00.3793940Z ##[endgroup] 2025-12-04T09:58:00.3823232Z + [[ -n '' ]] 2025-12-04T09:58:00.3824279Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-12-04T09:58:01.3740364Z /home/ec2-user/.local/lib/python3.9/site-packages/torch/_subclasses/functional_tensor.py:276: UserWarning: Failed to initialize NumPy: No module named 'numpy' (Triggered internally at /pytorch/torch/csrc/utils/tensor_numpy.cpp:81.) 2025-12-04T09:58:01.3741537Z cpu = _conversion_method_template(device=torch.device("cpu")) 2025-12-04T09:58:02.3355391Z ##[group]Run set -eux 2025-12-04T09:58:02.3355666Z set -eux 2025-12-04T09:58:02.3355887Z  2025-12-04T09:58:02.3356126Z # TODO (huydhn): Implement this part 2025-12-04T09:58:02.3356505Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:58:02.3367133Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:02.3367502Z env: 2025-12-04T09:58:02.3367723Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:02.3367984Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:02.3368299Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:02.3368836Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:02.3369309Z DEVICE_NAME: 2025-12-04T09:58:02.3369532Z DEVICE_TYPE: 2025-12-04T09:58:02.3369752Z ##[endgroup] 2025-12-04T09:58:02.3400629Z + echo 'dependencies={}' 2025-12-04T09:58:02.3480679Z ##[group]Run set -eux 2025-12-04T09:58:02.3480954Z set -eux 2025-12-04T09:58:02.3481176Z  2025-12-04T09:58:02.3481405Z if [[ -n "" ]]; then 2025-12-04T09:58:02.3481662Z  source "" 2025-12-04T09:58:02.3481889Z fi 2025-12-04T09:58:02.3482094Z  2025-12-04T09:58:02.3482348Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-12-04T09:58:02.3482773Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-12-04T09:58:02.3483244Z  # We don't want the job to fail if the directory doesn't exist 2025-12-04T09:58:02.3483613Z  exit 0 2025-12-04T09:58:02.3483821Z fi 2025-12-04T09:58:02.3484031Z  2025-12-04T09:58:02.3484270Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-12-04T09:58:02.3484719Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-12-04T09:58:02.3485369Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-12-04T09:58:02.3485783Z  --metadata "${BENCHMARK_METADATA}" \ 2025-12-04T09:58:02.3486182Z  --runners "${RUNNER_INFO}" \ 2025-12-04T09:58:02.3486516Z  --dependencies "${DEPENDENCIES}" \ 2025-12-04T09:58:02.3486840Z  --dry-run 2025-12-04T09:58:02.3487158Z else 2025-12-04T09:58:02.3487518Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-12-04T09:58:02.3488046Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-12-04T09:58:02.3488454Z  --metadata "${BENCHMARK_METADATA}" \ 2025-12-04T09:58:02.3488798Z  --runners "${RUNNER_INFO}" \ 2025-12-04T09:58:02.3489128Z  --dependencies "${DEPENDENCIES}" 2025-12-04T09:58:02.3489444Z fi 2025-12-04T09:58:02.3498429Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:02.3498775Z env: 2025-12-04T09:58:02.3498986Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:02.3499251Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:02.3499553Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:02.3500091Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:02.3500559Z DEVICE_NAME: 2025-12-04T09:58:02.3500783Z DEVICE_TYPE: 2025-12-04T09:58:02.3501023Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-12-04T09:58:02.3501322Z DRY_RUN: false 2025-12-04T09:58:02.3502798Z BENCHMARK_METADATA: {"timestamp": 1764842280, "schema_version": "v3", "name": "linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32", "workflow_id": 19922826259, "run_attempt": 1, "job_id": 57118183167} 2025-12-04T09:58:02.3504790Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 16, "avail_mem_in_gb": 62, "extra_info": {"hostname": "ip-10-0-66-49.ec2.internal"}, "name": "cuda", "type": "NVIDIA A10G", "gpu_count": 1, "avail_gpu_mem_in_gb": 22}] 2025-12-04T09:58:02.3505698Z DEPENDENCIES: {} 2025-12-04T09:58:02.3505922Z ##[endgroup] 2025-12-04T09:58:02.3536604Z + [[ -n '' ]] 2025-12-04T09:58:02.3537010Z + [[ ! -d test/test-reports ]] 2025-12-04T09:58:02.3537399Z + [[ false == \t\r\u\e ]] 2025-12-04T09:58:02.3540473Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py --benchmark-results-dir test/test-reports --metadata '{"timestamp": 1764842280, "schema_version": "v3", "name": "linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32", "workflow_id": 19922826259, "run_attempt": 1, "job_id": 57118183167}' --runners '[{"cpu_info": "x86_64", "cpu_count": 16, "avail_mem_in_gb": 62, "extra_info": {"hostname": "ip-10-0-66-49.ec2.internal"}, "name": "cuda", "type": "NVIDIA A10G", "gpu_count": 1, "avail_gpu_mem_in_gb": 22}]' --dependencies '{}' 2025-12-04T09:58:02.5156123Z /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py:236: UserWarning: {'included': [{'test_file': 'inductor/test_aot_inductor'}, {'test_file': 'inductor/test_torchinductor'}, {'test_file': 'inductor/test_torchinductor_dynamic_shapes'}, {'test_file': 'inductor/test_torchinductor_codegen_dynamic_shapes'}, {'test_file': 'inductor/test_kernel_benchmark'}, {'test_file': 'inductor/test_torchinductor_opinfo'}, {'test_file': 'inductor/test_pattern_matcher'}, {'test_file': 'inductor/test_cuda_repro'}, {'test_file': 'inductor/test_cudagraph_trees'}, {'test_file': 'dynamo/test_activation_checkpointing'}, {'test_file': 'dynamo/test_logging'}, {'test_file': 'dynamo/test_repros'}, {'test_file': 'inductor/test_flex_attention'}, {'test_file': 'inductor/test_cuda_select_algorithm'}, {'test_file': 'inductor/test_compile_subprocess'}, {'test_file': 'inductor/test_flex_decoding'}, {'test_file': 'inductor/test_deterministic'}, {'test_file': 'export/test_retraceability'}, {'test_file': 'inductor/test_fp8'}, {'test_file': 'dynamo/test_model_output'}, {'test_file': 'inductor/test_triton_kernels'}, {'test_file': 'inductor/test_extension_backend'}, {'test_file': 'inductor/test_native_matmul'}, {'test_file': 'inductor/test_loop_ordering'}, {'test_file': 'export/test_serdes'}, {'test_file': 'dynamo/test_regional_inductor'}, {'test_file': 'dynamo/test_fx_graph_runnable'}, {'test_file': 'dynamo/test_backends'}, {'test_file': 'inductor/test_aot_inductor_package'}, {'test_file': 'inductor/test_decompose_mem_bound_mm'}, {'test_file': 'inductor/test_op_dtype_prop'}, {'test_file': 'inductor/test_online_softmax'}, {'test_file': 'inductor/test_memory'}, {'test_file': 'dynamo/test_streams'}, {'test_file': 'inductor/test_unbacked_symints'}, {'test_file': 'inductor/test_scatter_optimization'}, {'test_file': 'inductor/test_mix_order_reduction'}, {'test_file': 'inductor/test_padding'}, {'test_file': 'dynamo/test_aot_compile'}, {'test_file': 'dynamo/test_sets'}, {'test_file': 'dynamo/test_wrap_inductor_compiled_regions'}, {'test_file': 'dynamo/test_callback'}, {'test_file': 'dynamo/test_compiler_bisector'}, {'test_file': 'inductor/test_custom_op_autotune'}, {'test_file': 'inductor/test_cudagraph_trees_expandable_segments'}, {'test_file': 'dynamo/test_decorators'}, {'test_file': 'test_privateuseone_python_backend'}, {'test_file': 'inductor/test_collective_autotuning'}, {'test_file': 'test_varlen_attention'}, {'test_file': 'test_cuda'}, {'test_file': 'test_transformers'}, {'test_file': 'test_autograd'}, {'test_file': 'test_sparse'}, {'test_file': 'higher_order_ops/test_local_map'}, {'test_file': 'test_dataloader'}, {'test_file': 'higher_order_ops/test_invoke_subgraph'}, {'test_file': 'test_ci_sanity_check_fail'}, {'test_file': 'test_ops_fwd_gradients'}, {'test_file': 'test_ops_gradients'}, {'test_file': 'test_nestedtensor'}, {'test_file': 'test_linalg'}, {'test_file': 'test_cuda_expandable_segments'}, {'test_file': 'test_public_bindings'}, {'test_file': 'functorch/test_dims'}, {'test_file': 'test_sparse_csr'}, {'test_file': 'functorch/test_ops'}, {'test_file': 'functorch/test_vmap'}, {'test_file': 'test_overrides'}, {'test_file': 'test_torchfuzz_repros'}, {'test_file': 'inductor/test_group_batch_fusion'}, {'test_file': 'dynamo/test_dynamic_shapes'}, {'test_file': 'inductor/test_cpu_repro'}, {'test_file': 'dynamo/test_after_aot'}, {'test_file': 'inductor/test_snode_runtime'}, {'test_file': 'inductor/test_minifier'}, {'test_file': 'inductor/test_compiled_autograd'}, {'test_file': 'inductor/test_custom_lowering'}, {'test_file': 'inductor/test_perf'}, {'test_file': 'inductor/test_fused_attention'}, {'test_file': 'inductor/test_binary_folding'}, {'test_file': 'inductor/test_mkldnn_pattern_matcher'}, {'test_file': 'inductor/test_inductor_freezing'}, {'test_file': 'inductor/test_layout_optim'}, {'test_file': 'dynamo/test_unspec'}, {'test_file': 'dynamo/test_higher_order_ops'}, {'test_file': 'inductor/test_mmdecomp'}, {'test_file': 'dynamo/test_ctx_manager'}, {'test_file': 'dynamo/test_exc'}, {'test_file': 'dynamo/test_misc'}, {'test_file': 'inductor/test_cpu_select_algorithm'}, {'test_file': 'inductor/test_aot_inductor_arrayref'}, {'test_file': 'inductor/test_cpu_cpp_wrapper'}, {'test_file': 'inductor/test_triton_cpu_backend'}, {'test_file': 'inductor/test_torchinductor_strided_blocks'}, {'test_file': 'test_custom_ops'}, {'test_file': 'test_content_store'}, {'test_file': 'inductor/test_halide'}, {'test_file': 'inductor/test_multi_kernel'}, {'test_file': 'inductor/test_analysis'}, {'test_file': 'inductor/test_pad_mm'}, {'test_file': 'inductor/test_triton_syntax'}, {'test_file': 'inductor/test_triton_extension_backend'}, {'test_file': 'test_sparse_semi_structured'}, {'test_file': 'inductor/test_op_completeness'}, {'test_file': 'inductor/test_subgraph_choice'}, {'test_file': 'inductor/test_b2b_gemm'}, {'test_file': 'inductor/test_triton_heuristics'}, {'test_file': 'inductor/test_cutedsl_grouped_mm'}, {'test_file': 'inductor/test_cpp_wrapper_hipify'}, {'test_file': 'inductor/test_ck_backend'}, {'test_file': 'inductor/test_inductor_utils'}, {'test_file': 'inductor/test_template_heuristics_registry'}, {'test_file': 'inductor/test_async_compile'}, {'test_file': 'inductor/test_gpu_cpp_wrapper'}, {'test_file': 'export/test_export_training_ir_to_run_decomp'}, {'test_file': 'dynamo/test_deque_reconstruct'}, {'test_file': 'inductor/test_utils'}, {'test_file': 'inductor/test_indexing'}, {'test_file': 'inductor/test_inductor_annotations'}, {'test_file': 'inductor/test_compile_worker'}, {'test_file': 'dynamo/test_einops'}, {'test_file': 'inductor/test_external_callables'}, {'test_file': 'test_testing'}, {'test_file': 'dynamo/test_fx_passes_pre_grad'}, {'test_file': 'inductor/test_autoheuristic'}, {'test_file': 'export/test_strict_export_v2'}, {'test_file': 'inductor/test_flex_flash'}, {'test_file': 'inductor/test_segmented_tree'}, {'test_file': 'inductor/test_kernel_optimization'}, {'test_file': 'inductor/test_metrics'}, {'test_file': 'export/test_unflatten_training_ir'}, {'test_file': 'inductor/test_fx_fusion'}, {'test_file': 'inductor/test_xpu_basic'}, {'test_file': 'dynamo/test_inline_and_install'}, {'test_file': 'export/test_functionalized_assertions'}, {'test_file': 'inductor/test_selective_lowering'}, {'test_file': 'dynamo/test_base_output'}, {'test_file': 'inductor/test_lookup_table'}, {'test_file': 'inductor/test_cooperative_reductions'}, {'test_file': 'export/test_serialize'}, {'test_file': 'inductor/test_cutedsl_template'}, {'test_file': 'inductor/test_benchmark_fusion'}, {'test_file': 'inductor/test_inductor_scheduler'}, {'test_file': 'inductor/test_move_constructors_to_gpu'}, {'test_file': 'export/test_export_strict'}, {'test_file': 'dynamo/test_modules'}, {'test_file': 'inductor/test_remote_cache'}, {'test_file': 'inductor/test_coordinate_descent_tuner'}, {'test_file': 'inductor/test_inplace_padding'}, {'test_file': 'inductor/test_cudacodecache'}, {'test_file': 'inductor/test_minifier_utils'}, {'test_file': 'inductor/test_debug_trace'}, {'test_file': 'dynamo/test_recompiles'}, {'test_file': 'inductor/test_foreach'}, {'test_file': 'export/test_tree_utils'}, {'test_file': 'inductor/test_triton_wrapper'}, {'test_file': 'inductor/test_static_cuda_launcher'}, {'test_file': 'export/test_dynamic_shapes'}, {'test_file': 'dynamo/test_sdpa'}, {'test_file': 'dynamo/test_utils'}, {'test_file': 'inductor/test_provenance_tracing'}, {'test_file': 'inductor/test_combo_kernels'}, {'test_file': 'inductor/test_codegen_triton'}, {'test_file': 'dynamo/test_frame_init'}, {'test_file': 'inductor/test_device_assert'}, {'test_file': 'dynamo/test_skip_non_tensor'}, {'test_file': 'dynamo/test_skip_guard_eval_unsafe'}, {'test_file': 'dynamo/test_interop'}, {'test_file': 'functorch/test_eager_transforms'}, {'test_file': 'inductor/test_control_deps'}, {'test_file': 'inductor/test_benchmarking'}, {'test_file': 'inductor/test_helion_kernels'}, {'test_file': 'inductor/test_quantization'}, {'test_file': 'inductor/test_best_config'}, {'test_file': 'export/test_tools'}, {'test_file': 'dynamo/test_buffers_override'}, {'test_file': 'inductor/test_inplacing_pass'}, {'test_file': 'inductor/test_aot_inductor_custom_ops'}, {'test_file': 'inductor/test_split_cat_fx_passes'}, {'test_file': 'inductor/test_profiler'}, {'test_file': 'inductor/test_memory_planning'}, {'test_file': 'inductor/test_mem_estimation'}, {'test_file': 'dynamo/test_view'}, {'test_file': 'inductor/test_cutlass_evt'}, {'test_file': 'dynamo/test_reconstruct'}, {'test_file': 'dynamo/test_aot_autograd'}, {'test_file': 'export/test_cpp_serdes'}, {'test_file': 'inductor/test_cache'}, {'test_file': 'inductor/test_block_analysis'}, {'test_file': 'dynamo/test_subgraphs'}, {'test_file': 'dynamo/test_pre_dispatch'}, {'test_file': 'inductor/test_custom_post_grad_passes'}, {'test_file': 'dynamo/test_fx_annotate'}, {'test_file': 'dynamo/test_pgo'}, {'test_file': 'dynamo/test_config'}, {'test_file': 'dynamo/test_metrics_context'}, {'test_file': 'export/test_package'}, {'test_file': 'export/test_export_opinfo'}, {'test_file': 'dynamo/test_nops'}, {'test_file': 'inductor/test_graph_transform_observer'}, {'test_file': 'inductor/test_aot_inductor_utils'}, {'test_file': 'export/test_db'}, {'test_file': 'dynamo/test_export_mutations'}, {'test_file': 'inductor/test_config'}, {'test_file': 'inductor/test_dependencies'}, {'test_file': 'inductor/test_fuzzer'}, {'test_file': 'dynamo/test_global'}, {'test_file': 'inductor/test_control_flow'}, {'test_file': 'dynamo/test_graph_region_tracker'}, {'test_file': 'dynamo/test_unittest'}, {'test_file': 'inductor/test_compile'}, {'test_file': 'dynamo/test_functions'}, {'test_file': 'inductor/test_ordered_set'}, {'test_file': 'inductor/test_pallas'}, {'test_file': 'dynamo/test_install_free_tensors'}, {'test_file': 'inductor/test_torchinductor_codegen_config_overrides'}, {'test_file': 'export/test_passes'}, {'test_file': 'dynamo/test_autograd_function'}, {'test_file': 'inductor/test_codecache'}, {'test_file': 'dynamo/test_cudagraphs'}, {'test_file': 'inductor/test_alignment'}, {'test_file': 'dynamo/test_profiler'}, {'test_file': 'dynamo/test_guard_serialization'}, {'test_file': 'dynamo/test_compile'}, {'test_file': 'dynamo/test_nested_graph_breaks'}, {'test_file': 'dynamo/test_dicts'}, {'test_file': 'inductor/test_needs_exact_strides'}, {'test_file': 'inductor/test_auto_functionalize'}, {'test_file': 'inductor/test_split_cat_fx_aten_passes'}, {'test_file': 'inductor/test_minifier_isolate'}, {'test_file': 'dynamo/test_list'}, {'test_file': 'dynamo/test_resume'}, {'test_file': 'inductor/test_augmented_graph_helper'}, {'test_file': 'dynamo/test_deviceguard'}, {'test_file': 'dynamo/test_sources'}, {'test_file': 'dynamo/test_backward_higher_order_ops'}, {'test_file': 'dynamo/test_modes'}, {'test_file': 'dynamo/test_optimizers'}, {'test_file': 'export/test_torchbind'}, {'test_file': 'inductor/test_custom_partitioner_fn'}, {'test_file': 'dynamo/test_debug_utils'}, {'test_file': 'dynamo/test_base_hop'}, {'test_file': 'dynamo/test_export'}, {'test_file': 'dynamo/test_package'}, {'test_file': 'inductor/test_efficient_conv_bn_eval'}, {'test_file': 'inductor/test_torchbind'}, {'test_file': 'dynamo/test_python_dispatcher'}, {'test_file': 'export/test_swap'}, {'test_file': 'export/test_unflatten'}, {'test_file': 'dynamo/test_verify_correctness'}, {'test_file': 'inductor/test_fxir_backend'}, {'test_file': 'dynamo/test_cudagraphs_expandable_segments'}, {'test_file': 'inductor/test_caching'}, {'test_file': 'dynamo/test_aot_autograd_cache'}, {'test_file': 'dynamo/test_flat_apply'}, {'test_file': 'dynamo/test_input_attr_tracking'}, {'test_file': 'dynamo/test_graph_deduplication'}, {'test_file': 'inductor/test_distributed_patterns'}, {'test_file': 'dynamo/test_structured_trace'}, {'test_file': 'dynamo/test_error_messages'}, {'test_file': 'dynamo/test_bytecode_utils'}, {'test_file': 'dynamo/test_fake_distributed'}, {'test_file': 'inductor/test_mps_basic'}, {'test_file': 'export/test_nativert'}, {'test_file': 'export/test_hop'}, {'test_file': 'dynamo/test_tree_map'}, {'test_file': 'dynamo/test_minifier'}, {'test_file': 'dynamo/test_guard_manager'}, {'test_file': 'export/test_schema'}, {'test_file': 'dynamo/test_torchrec'}, {'test_file': 'export/test_pass_infra'}, {'test_file': 'export/test_experimental'}, {'test_file': 'export/test_converter'}, {'test_file': 'export/test_export'}, {'test_file': 'test_model_exports_to_core_aten'}, {'test_file': 'dynamo/test_precompile_context'}, {'test_file': 'dynamo/test_trace_rules'}, {'test_file': 'export/test_upgrader'}, {'test_file': 'dynamo/test_hooks'}, {'test_file': 'dynamo/test_reorder_logs'}, {'test_file': 'dynamo/test_subclasses'}, {'test_file': 'dynamo/test_exceptions'}, {'test_file': 'dynamo/test_generator'}, {'test_file': 'export/test_lift_unlift'}, {'test_file': 'export/test_verifier'}, {'test_file': 'export/test_sparse'}, {'test_file': 'dynamo/test_python_autograd'}, {'test_file': 'export/test_draft_export'}, {'test_file': 'dynamo/test_comptime'}, {'test_file': 'test_sort_and_select'}, {'test_file': 'functorch/test_rearrange'}, {'test_file': 'functorch/test_parsing'}, {'test_file': 'test_package'}, {'test_file': 'profiler/test_profiler'}, {'test_file': 'test_mkl_verbose'}, {'test_file': 'test_comparison_utils'}, {'test_file': 'functorch/test_ac_logging'}, {'test_file': 'test_mkldnn_verbose'}, {'test_file': 'test_cpp_api_parity'}, {'test_file': 'test_utils_config_module'}, {'test_file': 'test_hop_infra'}, {'test_file': 'test_appending_byte_serializer'}, {'test_file': 'test_license'}, {'test_file': 'test_ao_sparsity'}, {'test_file': 'test_autoload'}, {'test_file': 'nn/attention/test_open_registry'}, {'test_file': 'xpu/test_fusion'}, {'test_file': 'test_as_strided'}, {'test_file': 'test_foreach'}, {'test_file': 'test_proxy_tensor'}, {'test_file': 'torch_np/test_binary_ufuncs'}, {'test_file': 'torch_np/test_unary_ufuncs'}, {'test_file': 'test_utils_filelock'}, {'test_file': 'test_extension_utils'}, {'test_file': 'test_rename_privateuse1_to_existing_device'}, {'test_file': 'nn/attention/test_fa4'}, {'test_file': 'typing/test_python_operators'}, {'test_file': 'test_functionalization'}, {'test_file': 'torch_np/test_dtype'}, {'test_file': 'test_file_check'}, {'test_file': 'profiler/test_kineto'}, {'test_file': 'test_flop_counter'}, {'test_file': 'backends/xeon/test_launch'}, {'test_file': 'test_show_pickle'}, {'test_file': 'test_openmp'}, {'test_file': 'test_expanded_weights'}, {'test_file': 'test_module_tracker'}, {'test_file': 'torch_np/numpy_tests/core/test_scalarinherit'}, {'test_file': 'test_tensorexpr_pybind'}, {'test_file': 'test_fx_experimental'}, {'test_file': 'functorch/test_ac_knapsack'}, {'test_file': 'torch_np/test_nep50_examples'}, {'test_file': 'test_torch'}, {'test_file': 'xpu/test_gemm'}, {'test_file': 'test_fx_passes'}, {'test_file': 'functorch/test_logging'}, {'test_file': 'test_namedtensor'}, {'test_file': 'test_tensorexpr'}, {'test_file': 'functorch/test_minifier'}, {'test_file': 'higher_order_ops/test_invoke_quant'}, {'test_file': 'torch_np/test_basic'}, {'test_file': 'test_jiterator'}, {'test_file': 'test_native_functions'}, {'test_file': 'test_typing'}, {'test_file': 'higher_order_ops/test_with_effects'}, {'test_file': 'test_weak'}, {'test_file': 'test_complex'}, {'test_file': 'test_optim'}, {'test_file': 'lazy/test_functionalization'}, {'test_file': 'torch_np/test_random'}, {'test_file': 'nn/test_multihead_attention'}, {'test_file': 'test_legacy_vmap'}, {'test_file': 'lazy/test_bindings'}, {'test_file': 'xpu/test_conv'}, {'test_file': 'test_utils'}, {'test_file': 'test_pytree'}, {'test_file': 'test_namedtuple_return_api'}, {'test_file': 'profiler/test_record_function'}, {'test_file': 'test_compile_benchmark_util'}, {'test_file': 'test_set_default_mobile_cpu_allocator'}, {'test_file': 'test_fake_tensor'}, {'test_file': 'test_stateless'}, {'test_file': 'functorch/test_ac'}, {'test_file': 'test_binary_ufuncs'}, {'test_file': 'higher_order_ops/test_print'}, {'test_file': 'test_per_overload_api'}, {'test_file': 'torch_np/numpy_tests/core/test_einsum'}, {'test_file': 'test_multiprocessing'}, {'test_file': 'test_out_dtype_op'}, {'test_file': 'torch_np/test_ufuncs_basic'}, {'test_file': 'lazy/test_step_closures'}, {'test_file': 'functorch/dim/test_getsetitem'}, {'test_file': 'test_numpy_interop'}, {'test_file': 'profiler/test_cpp_thread'}, {'test_file': 'test_segment_reductions'}, {'test_file': 'test_opaque_obj_v2'}, {'test_file': 'test_autograd_fallback'}, {'test_file': 'test_type_hints'}, {'test_file': 'functorch/test_aot_joint_with_descriptors'}, {'test_file': 'test_functionalization_of_rng_ops'}, {'test_file': 'test_fx_reinplace_pass'}, {'test_file': 'functorch/test_control_flow'}, {'test_file': 'test_modules'}, {'test_file': 'nn/test_packed_sequence'}, {'test_file': 'test_numa_binding'}, {'test_file': 'test_pruning_op'}, {'test_file': 'test_jit_fuser_te'}, {'test_file': 'test_autocast'}, {'test_file': 'test_logging'}, {'test_file': 'test_python_dispatch'}, {'test_file': 'nn/test_lazy_modules'}, {'test_file': 'nn/test_pruning'}, {'test_file': 'test_monitor'}, {'test_file': 'test_cuda_sanitizer'}, {'test_file': 'test_bundled_inputs'}, {'test_file': 'torch_np/numpy_tests/core/test_numeric'}, {'test_file': 'torch_np/numpy_tests/core/test_multiarray'}, {'test_file': 'test_itt'}, {'test_file': 'torch_np/numpy_tests/lib/test_function_base'}, {'test_file': 'test_masked'}, {'test_file': 'test_sympy_utils'}, {'test_file': 'test_jit_disabled'}, {'test_file': 'test_subclass'}, {'test_file': 'test_import_stats'}, {'test_file': 'functorch/test_vmap_registrations'}, {'test_file': 'nn/test_parametrization'}, {'test_file': 'test_mkldnn_fusion'}, {'test_file': 'test_cpp_extensions_mtia_backend'}, {'test_file': 'lazy/test_ts_opinfo'}, {'test_file': 'test_dynamic_shapes'}, {'test_file': 'complex_tensor/test_complex_tensor'}, {'test_file': 'optim/test_lrscheduler'}, {'test_file': 'optim/test_swa_utils'}, {'test_file': 'cpp_extensions/python_agnostic_extension/test/test_python_agnostic'}, {'test_file': 'functorch/test_memory_efficient_fusion'}, {'test_file': 'torch_np/numpy_tests/lib/test_histograms'}, {'test_file': 'torch_np/test_indexing'}, {'test_file': 'test_schema_check'}, {'test_file': 'test_tensorboard'}, {'test_file': 'torch_np/numpy_tests/core/test_indexing'}, {'test_file': 'test_futures'}, {'test_file': 'test_tensor_creation_ops'}, {'test_file': 'nn/test_dropout'}, {'test_file': 'functorch/dim/test_split'}, {'test_file': 'torch_np/numpy_tests/lib/test_type_check'}, {'test_file': 'cpp_extensions/test_libtorch_agnostic'}, {'test_file': 'test_cpp_extensions_stream_and_event'}, {'test_file': 'profiler/test_execution_trace'}, {'test_file': 'test_dispatch'}, {'test_file': 'test_datapipe'}, {'test_file': 'test_numba_integration'}, {'test_file': 'test_functional_optim'}, {'test_file': 'test_maskedtensor'}, {'test_file': 'benchmark_utils/test_benchmark_utils'}, {'test_file': 'torch_np/numpy_tests/core/test_scalarmath'}, {'test_file': 'test_scaled_matmul_cuda'}, {'test_file': 'torch_np/numpy_tests/core/test_shape_base'}, {'test_file': 'test_vulkan'}, {'test_file': 'lazy/test_generator'}, {'test_file': 'nn/test_convolution'}, {'test_file': 'torch_np/numpy_tests/linalg/test_linalg'}, {'test_file': 'torch_np/numpy_tests/core/test_dtype'}, {'test_file': 'lazy/test_debug_util'}, {'test_file': 'nn/test_load_state_dict'}, {'test_file': 'test_shape_ops'}, {'test_file': 'nn/test_module_hooks'}, {'test_file': 'torch_np/numpy_tests/lib/test_twodim_base'}, {'test_file': 'profiler/test_memory_profiler'}, {'test_file': 'test_jit_llga_fuser'}, {'test_file': 'test_serialization'}, {'test_file': 'optim/test_optim'}, {'test_file': 'test_indexing'}, {'test_file': 'torch_np/numpy_tests/fft/test_pocketfft'}, {'test_file': 'torch_np/numpy_tests/lib/test_shape_base_'}, {'test_file': 'torch_np/numpy_tests/core/test_getlimits'}, {'test_file': 'torch_np/test_ndarray_methods'}, {'test_file': 'test_view_ops'}, {'test_file': 'test_type_info'}, {'test_file': 'functorch/test_aotdispatch'}, {'test_file': 'test_nn'}, {'test_file': 'torch_np/numpy_tests/core/test_dlpack'}, {'test_file': 'test_multiprocessing_spawn'}, {'test_file': 'test_scatter_gather_ops'}, {'test_file': 'test_cuda_multigpu'}, {'test_file': 'test_mkldnn'}, {'test_file': 'torch_np/numpy_tests/lib/test_index_tricks'}, {'test_file': 'test_jit_autocast'}, {'test_file': 'nn/test_pooling'}, {'test_file': 'nn/test_embedding'}, {'test_file': 'test_unary_ufuncs'}, {'test_file': 'test_xnnpack_integration'}, {'test_file': 'test_cuda_trace'}, {'test_file': 'test_native_mha'}, {'test_file': 'torch_np/numpy_tests/core/test_numerictypes'}, {'test_file': 'test_cuda_nvml_based_avail'}, {'test_file': 'test_function_schema'}, {'test_file': 'test_accelerator'}, {'test_file': 'nn/test_init'}, {'test_file': 'torch_np/numpy_tests/core/test_scalar_methods'}, {'test_file': 'torch_np/numpy_tests/fft/test_helper'}, {'test_file': 'test_mobile_optimizer'}, {'test_file': 'torch_np/test_function_base'}, {'test_file': 'test_type_promotion'}, {'test_file': 'torch_np/test_scalars_0D_arrays'}, {'test_file': 'test_cuda_primary_ctx'}, {'test_file': 'profiler/test_profiler_tree'}, {'test_file': 'torch_np/numpy_tests/lib/test_arraysetops'}, {'test_file': 'test_dlpack'}, {'test_file': 'profiler/test_torch_tidy'}, {'test_file': 'lazy/test_reuse_ir'}, {'test_file': 'test_functional_autograd_benchmark'}, {'test_file': 'test_reductions'}, {'test_file': 'torch_np/test_reductions'}, {'test_file': 'torch_np/numpy_tests/core/test_scalar_ctors'}, {'test_file': 'torch_np/numpy_tests/lib/test_arraypad'}, {'test_file': 'test_prims'}, {'test_file': 'test_spectral_ops'}, {'test_file': 'profiler/test_python_tracer'}, {'test_file': 'cpp_extensions/libtorch_agnostic_2_10_extension/test_version_compatibility'}, {'test_file': 'distributions/test_distributions'}, {'test_file': 'test_autoload_disable'}, {'test_file': 'test_autoload_enable'}, {'test_file': 'test_cpp_extensions_aot_ninja'}, {'test_file': 'test_cpp_extensions_aot_no_ninja'}], 'excluded': []} from test/test-reports/td_exclusions-b4040f7bcae1d1491729.json is not a benchmark record, skipping 2025-12-04T09:58:02.5222335Z warn(f"{result} from {filepath} is not a benchmark record, skipping") 2025-12-04T09:58:02.5223893Z /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py:236: UserWarning: {'included': [{'test_file': 'cpp/test_jit'}, {'test_file': 'cpp/test_lazy'}], 'excluded': []} from test/test-reports/td_exclusions-faa19e9dd28552f11069.json is not a benchmark record, skipping 2025-12-04T09:58:02.5225449Z warn(f"{result} from {filepath} is not a benchmark record, skipping") 2025-12-04T09:58:02.5338739Z ##[group]Run cat test/**/*_toprint.log || true 2025-12-04T09:58:02.5339119Z cat test/**/*_toprint.log || true 2025-12-04T09:58:02.5348169Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:02.5348524Z env: 2025-12-04T09:58:02.5348755Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:02.5349011Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:02.5349321Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:02.5349851Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:02.5350316Z DEVICE_NAME: 2025-12-04T09:58:02.5350535Z DEVICE_TYPE: 2025-12-04T09:58:02.5350740Z ##[endgroup] 2025-12-04T09:58:02.5465145Z Test results will be stored in test-reports/python-pytest/test_ci_sanity_check_fail/test_ci_sanity_check_fail-024947735ba5c526.xml 2025-12-04T09:58:02.5465838Z ============================= test session starts ============================== 2025-12-04T09:58:02.5466414Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:58:02.5466899Z cachedir: .pytest_cache 2025-12-04T09:58:02.5467484Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:58:02.5468130Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:58:02.5468600Z configfile: pytest.ini 2025-12-04T09:58:02.5469201Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:58:02.5470033Z collecting ... collected 2 items 2025-12-04T09:58:02.5470387Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:58:02.5470742Z Running 0 items in this shard: 2025-12-04T09:58:02.5470931Z 2025-12-04T09:58:02.5471490Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ci_sanity_check_fail/test_ci_sanity_check_fail-024947735ba5c526.xml - 2025-12-04T09:58:02.5472285Z ============================ no tests ran in 0.01s ============================= 2025-12-04T09:58:02.5523983Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-12-04T09:58:02.5524340Z kill "$MONITOR_SCRIPT_PID" 2025-12-04T09:58:02.5533448Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:02.5533898Z env: 2025-12-04T09:58:02.5534114Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:02.5534377Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:02.5534678Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:02.5535200Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:02.5536026Z DEVICE_NAME: 2025-12-04T09:58:02.5536248Z DEVICE_TYPE: 2025-12-04T09:58:02.5536474Z MONITOR_SCRIPT_PID: 59432 2025-12-04T09:58:02.5536727Z ##[endgroup] 2025-12-04T09:58:02.5566604Z /home/ec2-user/actions-runner/_work/_temp/1461f05a-1c94-4f9d-8112-21085d64d569.sh: line 1: kill: (59432) - No such process 2025-12-04T09:58:02.5580414Z ##[error]Process completed with exit code 1. 2025-12-04T09:58:02.5718407Z Prepare all required actions 2025-12-04T09:58:02.5718804Z Getting action download info 2025-12-04T09:58:02.7718317Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T09:58:03.4327860Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-12-04T09:58:05.3895620Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-12-04T09:58:05.3895960Z with: 2025-12-04T09:58:05.3896313Z file-suffix: test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167 2025-12-04T09:58:05.3896803Z s3-bucket: gha-artifacts 2025-12-04T09:58:05.3897050Z env: 2025-12-04T09:58:05.3897270Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:05.3897534Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:05.3897846Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:05.3898378Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:05.3898886Z DEVICE_NAME: 2025-12-04T09:58:05.3899103Z DEVICE_TYPE: 2025-12-04T09:58:05.3899326Z ##[endgroup] 2025-12-04T09:58:05.4019858Z ##[group]Run # Remove any previous test jsons if they exist 2025-12-04T09:58:05.4020284Z # Remove any previous test jsons if they exist 2025-12-04T09:58:05.4020647Z rm -f test-jsons-*.zip 2025-12-04T09:58:05.4021072Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-12-04T09:58:05.4030241Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:05.4030670Z env: 2025-12-04T09:58:05.4030918Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:05.4031182Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:05.4031484Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:05.4032014Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:05.4032477Z DEVICE_NAME: 2025-12-04T09:58:05.4032689Z DEVICE_TYPE: 2025-12-04T09:58:05.4033045Z FILE_SUFFIX: test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167 2025-12-04T09:58:05.4033450Z ##[endgroup] 2025-12-04T09:58:05.5002915Z adding: test/test-reports/td_exclusions-b4040f7bcae1d1491729.json (deflated 82%) 2025-12-04T09:58:05.5025255Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-7db6a6e521716bbc.json (deflated 99%) 2025-12-04T09:58:05.5026636Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_codegen_dynamic_shapes/inductor.test_torchinductor_codegen_dynamic_shapes-667397d8b0e0d249.json (stored 0%) 2025-12-04T09:58:05.5027997Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-274766fd55a1aebe.json (deflated 98%) 2025-12-04T09:58:05.5031088Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-deb335f8881dbf3d.json (deflated 99%) 2025-12-04T09:58:05.5068517Z adding: test/test-reports/python-pytest/inductor.test_flex_attention/inductor.test_flex_attention-92d0889cf260b412.json (deflated 99%) 2025-12-04T09:58:05.5069489Z adding: test/test-reports/python-pytest/inductor.test_fp8/inductor.test_fp8-944065b13a109cb5.json (stored 0%) 2025-12-04T09:58:05.5070566Z adding: test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-55abd8e6b4532ae9.json (deflated 98%) 2025-12-04T09:58:05.5084237Z adding: test/test-reports/python-pytest/inductor.test_triton_kernels/inductor.test_triton_kernels-68c9c700698bff68.json (deflated 99%) 2025-12-04T09:58:05.5085902Z adding: test/test-reports/python-pytest/inductor.test_loop_ordering/inductor.test_loop_ordering-5ee2a3a0a4d76c5b.json (deflated 98%) 2025-12-04T09:58:05.5087553Z adding: test/test-reports/python-pytest/export.test_serdes/export.test_serdes-683bede13c6be09d.json (stored 0%) 2025-12-04T09:58:05.5089922Z adding: test/test-reports/python-pytest/inductor.test_scatter_optimization/inductor.test_scatter_optimization-f067953bd7ad79f7.json (stored 0%) 2025-12-04T09:58:05.5091967Z adding: test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-044d18a3f3486007.json (stored 0%) 2025-12-04T09:58:05.5093757Z adding: test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-3a59a6b9d0aa5ff8.json (stored 0%) 2025-12-04T09:58:05.5095742Z adding: test/test-reports/python-pytest/inductor.test_custom_op_autotune/inductor.test_custom_op_autotune-8053d2589810df17.json (stored 0%) 2025-12-04T09:58:05.5110429Z adding: test/test-reports/python-pytest/test_cuda/test_cuda-be50f111eb28e9fa.json (deflated 99%) 2025-12-04T09:58:05.5111173Z adding: test/test-reports/python-pytest/test_sparse/test_sparse-1f8fd945d58b94e8.json (stored 0%) 2025-12-04T09:58:05.5112010Z adding: test/test-reports/python-pytest/test_ci_sanity_check_fail/test_ci_sanity_check_fail-024947735ba5c526.json (stored 0%) 2025-12-04T09:58:05.5112922Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-bce2e111090231c8.json (stored 0%) 2025-12-04T09:58:05.5113800Z adding: test/test-reports/python-pytest/test_ops_gradients/test_ops_gradients-d325fd253497d5cc.json (stored 0%) 2025-12-04T09:58:05.5114643Z adding: test/test-reports/python-pytest/test_ops_gradients/test_ops_gradients-192c0166610ef1c8.json (stored 0%) 2025-12-04T09:58:05.5115499Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-423c2c6c6eabe2c7.json (stored 0%) 2025-12-04T09:58:05.5116374Z adding: test/test-reports/python-pytest/dynamo.test_after_aot/dynamo.test_after_aot-a91b437d94dc7298.json (stored 0%) 2025-12-04T09:58:05.5117326Z adding: test/test-reports/python-pytest/inductor.test_snode_runtime/inductor.test_snode_runtime-1e765d10347efbf7.json (stored 0%) 2025-12-04T09:58:05.5121667Z adding: test/test-reports/python-pytest/inductor.test_compiled_autograd/inductor.test_compiled_autograd-857de25fa6c3d7b5.json (deflated 98%) 2025-12-04T09:58:05.5124535Z adding: test/test-reports/python-pytest/test_testing/test_testing-509238380d0e9f60.json (deflated 99%) 2025-12-04T09:58:05.5125449Z adding: test/test-reports/python-pytest/inductor.test_autoheuristic/inductor.test_autoheuristic-b1e93791ac32a57e.json (stored 0%) 2025-12-04T09:58:05.5126509Z adding: test/test-reports/python-pytest/inductor.test_cutedsl_template/inductor.test_cutedsl_template-47a7a96144df1709.json (stored 0%) 2025-12-04T09:58:05.5498922Z adding: test/test-reports/python-pytest/inductor.test_benchmark_fusion/inductor.test_benchmark_fusion-1a5c9a2f1b3021da.json (deflated 94%) 2025-12-04T09:58:05.5499984Z adding: test/test-reports/python-pytest/inductor.test_remote_cache/inductor.test_remote_cache-e53c2a7e4fbe3b3f.json (stored 0%) 2025-12-04T09:58:05.5501081Z adding: test/test-reports/python-pytest/inductor.test_coordinate_descent_tuner/inductor.test_coordinate_descent_tuner-9ed05914eb073327.json (stored 0%) 2025-12-04T09:58:05.5584135Z adding: test/test-reports/python-pytest/inductor.test_inplace_padding/inductor.test_inplace_padding-ec8a12f82b4c1954.json (deflated 99%) 2025-12-04T09:58:05.5586678Z adding: test/test-reports/python-pytest/inductor.test_cudacodecache/inductor.test_cudacodecache-c55849964bd60b32.json (deflated 98%) 2025-12-04T09:58:05.5587725Z adding: test/test-reports/python-pytest/inductor.test_minifier_utils/inductor.test_minifier_utils-574f06f5728a4f68.json (stored 0%) 2025-12-04T09:58:05.5588856Z adding: test/test-reports/python-pytest/inductor.test_debug_trace/inductor.test_debug_trace-ffbc75868f825ab3.json (stored 0%) 2025-12-04T09:58:05.5589818Z adding: test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-7da9dd1b94d7565c.json (stored 0%) 2025-12-04T09:58:05.5590794Z adding: test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-0b0941e28d632943.json (stored 0%) 2025-12-04T09:58:05.5591987Z adding: test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-0c2b4708e9bf14de.json (stored 0%) 2025-12-04T09:58:05.5593262Z adding: test/test-reports/python-pytest/inductor.test_provenance_tracing/inductor.test_provenance_tracing-e8a06af29d1ce55a.json (stored 0%) 2025-12-04T09:58:05.5634016Z adding: test/test-reports/python-pytest/inductor.test_memory_planning/inductor.test_memory_planning-7b0b7b5cd4d35939.json (deflated 98%) 2025-12-04T09:58:05.5635024Z adding: test/test-reports/python-pytest/export.test_cpp_serdes/export.test_cpp_serdes-28f96f522b58f753.json (stored 0%) 2025-12-04T09:58:05.5636189Z adding: test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-22c864d375691836.json (stored 0%) 2025-12-04T09:58:05.5637165Z adding: test/test-reports/python-pytest/test_sort_and_select/test_sort_and_select-900006a153e58acf.json (stored 0%) 2025-12-04T09:58:05.5638093Z adding: test/test-reports/python-pytest/functorch.test_rearrange/functorch.test_rearrange-91f40a7195b2f228.json (stored 0%) 2025-12-04T09:58:05.5639062Z adding: test/test-reports/python-pytest/test_package/test_package-4f9aecb7033550fa.json (deflated 98%) 2025-12-04T09:58:05.5639913Z adding: test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-7e432f71a73d0aad.json (stored 0%) 2025-12-04T09:58:05.5640801Z adding: test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-197d280bfbc0adc1.json (stored 0%) 2025-12-04T09:58:05.5641677Z adding: test/test-reports/python-pytest/test_hop_infra/test_hop_infra-3a90bb4fa75fbe13.json (stored 0%) 2025-12-04T09:58:05.5642599Z adding: test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-6e708074a2279afc.json (stored 0%) 2025-12-04T09:58:05.5643529Z adding: test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-51844dfab6f78c47.json (stored 0%) 2025-12-04T09:58:05.5644382Z adding: test/test-reports/python-pytest/test_extension_utils/test_extension_utils-9ca2aaa70629c80b.json (stored 0%) 2025-12-04T09:58:05.5645288Z adding: test/test-reports/python-pytest/nn.attention.test_fa4/nn.attention.test_fa4-d1bda12ba20d2196.json (stored 0%) 2025-12-04T09:58:05.5646247Z adding: test/test-reports/python-pytest/typing.test_python_operators/typing.test_python_operators-ba4b7d5e3b95dbfb.json (stored 0%) 2025-12-04T09:58:05.5647247Z adding: test/test-reports/python-pytest/torch_np.test_dtype/torch_np.test_dtype-58897eedc88355ca.json (stored 0%) 2025-12-04T09:58:05.5648188Z adding: test/test-reports/python-pytest/test_file_check/test_file_check-9ff5837f5f97472d.json (stored 0%) 2025-12-04T09:58:05.5649046Z adding: test/test-reports/python-pytest/profiler.test_kineto/profiler.test_kineto-81a44d1732bd85bc.json (stored 0%) 2025-12-04T09:58:05.5650004Z adding: test/test-reports/python-pytest/functorch.test_ac_knapsack/functorch.test_ac_knapsack-56b46ef5e2d4cc8d.json (stored 0%) 2025-12-04T09:58:05.5651000Z adding: test/test-reports/python-pytest/torch_np.test_nep50_examples/torch_np.test_nep50_examples-a5a9215260d58c1d.json (stored 0%) 2025-12-04T09:58:05.5651878Z adding: test/test-reports/python-pytest/test_torch/test_torch-3d54dab073791e89.json (deflated 98%) 2025-12-04T09:58:05.5652634Z adding: test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-b46323f728df1f56.json (stored 0%) 2025-12-04T09:58:05.5653448Z adding: test/test-reports/python-pytest/test_binary_ufuncs/test_binary_ufuncs-3281db6a17fd8e68.json (stored 0%) 2025-12-04T09:58:05.5654343Z adding: test/test-reports/python-pytest/test_modules/test_modules-0165bb40c77358d1.json (deflated 98%) 2025-12-04T09:58:05.5655331Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-696665a6c90efdc4.json (stored 0%) 2025-12-04T09:58:05.5656506Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-f968ba80b11f10a3.json (stored 0%) 2025-12-04T09:58:05.5657536Z adding: test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-953e7ce7fc6caa5f.json (stored 0%) 2025-12-04T09:58:05.5658539Z adding: test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-48f47027a0649f8b.json (stored 0%) 2025-12-04T09:58:05.5659387Z adding: test/test-reports/python-pytest/test_shape_ops/test_shape_ops-85cca46ea4083bfa.json (stored 0%) 2025-12-04T09:58:05.5660326Z adding: test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-b02e0fce99e01635.json (deflated 98%) 2025-12-04T09:58:05.5661258Z adding: test/test-reports/python-pytest/test_indexing/test_indexing-c851479f46f8a70b.json (stored 0%) 2025-12-04T09:58:05.5662033Z adding: test/test-reports/python-pytest/test_type_info/test_type_info-1c1348d8e4ba037e.json (stored 0%) 2025-12-04T09:58:05.5662946Z adding: test/test-reports/python-pytest/functorch.test_aotdispatch/functorch.test_aotdispatch-0f6db4e497a4ab5e.json (stored 0%) 2025-12-04T09:58:05.5663915Z adding: test/test-reports/python-pytest/test_scatter_gather_ops/test_scatter_gather_ops-96cbf94077b2f70a.json (stored 0%) 2025-12-04T09:58:05.5664818Z adding: test/test-reports/python-pytest/test_cuda_multigpu/test_cuda_multigpu-4895f15a6c4e5281.json (deflated 98%) 2025-12-04T09:58:05.5665842Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_index_tricks/torch_np.numpy_tests.lib.test_index_tricks-506b926e701f49cf.json (stored 0%) 2025-12-04T09:58:05.5666917Z adding: test/test-reports/python-pytest/test_jit_autocast/test_jit_autocast-94bb29b8a8a5a7c3.json (stored 0%) 2025-12-04T09:58:05.5667823Z adding: test/test-reports/python-pytest/test_xnnpack_integration/test_xnnpack_integration-802360b65a8d44de.json (stored 0%) 2025-12-04T09:58:05.5668677Z adding: test/test-reports/python-pytest/nn.test_init/nn.test_init-da81cdd3c2c0ce98.json (stored 0%) 2025-12-04T09:58:05.5669620Z adding: test/test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-5cb9f1a76518723d.json (deflated 99%) 2025-12-04T09:58:05.5671124Z adding: test/test-reports/python-pytest/test_type_promotion/test_type_promotion-365dfbbc94f879c8.json (stored 0%) 2025-12-04T09:58:05.5672179Z adding: test/test-reports/python-pytest/test_reductions/test_reductions-88d67cc570f30743.json (stored 0%) 2025-12-04T09:58:05.5673100Z adding: test/test-reports/python-unittest/test_autoload/TEST-TestDeviceBackendAutoload-20251204095539.json (deflated 38%) 2025-12-04T09:58:05.5673912Z adding: test/test-reports/td_exclusions-faa19e9dd28552f11069.json (deflated 34%) 2025-12-04T09:58:05.5723285Z ##[group]Run # Remove any previous test reports if they exist 2025-12-04T09:58:05.5723750Z # Remove any previous test reports if they exist 2025-12-04T09:58:05.5724123Z rm -f test-reports-*.zip 2025-12-04T09:58:05.5724583Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-12-04T09:58:05.5733723Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:05.5734081Z env: 2025-12-04T09:58:05.5734295Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:05.5734561Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:05.5734871Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:05.5735673Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:05.5736144Z DEVICE_NAME: 2025-12-04T09:58:05.5736492Z DEVICE_TYPE: 2025-12-04T09:58:05.5736872Z FILE_SUFFIX: test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167 2025-12-04T09:58:05.5737287Z ##[endgroup] 2025-12-04T09:58:05.5876112Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-7db6a6e521716bbc.xml (deflated 99%) 2025-12-04T09:58:05.5877336Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_codegen_dynamic_shapes/inductor.test_torchinductor_codegen_dynamic_shapes-667397d8b0e0d249.xml (deflated 28%) 2025-12-04T09:58:05.5878621Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-274766fd55a1aebe.xml (deflated 98%) 2025-12-04T09:58:05.5881158Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-deb335f8881dbf3d.xml (deflated 99%) 2025-12-04T09:58:05.5917177Z adding: test/test-reports/python-pytest/inductor.test_flex_attention/inductor.test_flex_attention-92d0889cf260b412.xml (deflated 99%) 2025-12-04T09:58:05.5918139Z adding: test/test-reports/python-pytest/inductor.test_fp8/inductor.test_fp8-944065b13a109cb5.xml (deflated 27%) 2025-12-04T09:58:05.5919064Z adding: test/test-reports/python-pytest/dynamo.test_model_output/dynamo.test_model_output-55abd8e6b4532ae9.xml (deflated 97%) 2025-12-04T09:58:05.5931460Z adding: test/test-reports/python-pytest/inductor.test_triton_kernels/inductor.test_triton_kernels-68c9c700698bff68.xml (deflated 99%) 2025-12-04T09:58:05.5932862Z adding: test/test-reports/python-pytest/inductor.test_loop_ordering/inductor.test_loop_ordering-5ee2a3a0a4d76c5b.xml (deflated 98%) 2025-12-04T09:58:05.5933828Z adding: test/test-reports/python-pytest/export.test_serdes/export.test_serdes-683bede13c6be09d.xml (deflated 28%) 2025-12-04T09:58:05.5934851Z adding: test/test-reports/python-pytest/inductor.test_scatter_optimization/inductor.test_scatter_optimization-f067953bd7ad79f7.xml (deflated 28%) 2025-12-04T09:58:05.5936117Z adding: test/test-reports/python-pytest/inductor.test_padding/inductor.test_padding-044d18a3f3486007.xml (deflated 28%) 2025-12-04T09:58:05.5937059Z adding: test/test-reports/python-pytest/dynamo.test_callback/dynamo.test_callback-3a59a6b9d0aa5ff8.xml (deflated 28%) 2025-12-04T09:58:05.5938070Z adding: test/test-reports/python-pytest/inductor.test_custom_op_autotune/inductor.test_custom_op_autotune-8053d2589810df17.xml (deflated 28%) 2025-12-04T09:58:05.5955179Z adding: test/test-reports/python-pytest/test_cuda/test_cuda-be50f111eb28e9fa.xml (deflated 99%) 2025-12-04T09:58:05.5955940Z adding: test/test-reports/python-pytest/test_sparse/test_sparse-1f8fd945d58b94e8.xml (deflated 28%) 2025-12-04T09:58:05.5957349Z adding: test/test-reports/python-pytest/test_ci_sanity_check_fail/test_ci_sanity_check_fail-024947735ba5c526.xml (deflated 27%) 2025-12-04T09:58:05.5959209Z adding: test/test-reports/python-pytest/test_ops_fwd_gradients/test_ops_fwd_gradients-bce2e111090231c8.xml (deflated 28%) 2025-12-04T09:58:05.5960991Z adding: test/test-reports/python-pytest/test_ops_gradients/test_ops_gradients-d325fd253497d5cc.xml (deflated 28%) 2025-12-04T09:58:05.5962925Z adding: test/test-reports/python-pytest/test_ops_gradients/test_ops_gradients-192c0166610ef1c8.xml (deflated 28%) 2025-12-04T09:58:05.5964661Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-423c2c6c6eabe2c7.xml (deflated 27%) 2025-12-04T09:58:05.5966442Z adding: test/test-reports/python-pytest/dynamo.test_after_aot/dynamo.test_after_aot-a91b437d94dc7298.xml (deflated 28%) 2025-12-04T09:58:05.5967559Z adding: test/test-reports/python-pytest/inductor.test_snode_runtime/inductor.test_snode_runtime-1e765d10347efbf7.xml (deflated 28%) 2025-12-04T09:58:05.5968637Z adding: test/test-reports/python-pytest/inductor.test_compiled_autograd/inductor.test_compiled_autograd-857de25fa6c3d7b5.xml (deflated 99%) 2025-12-04T09:58:05.5969580Z adding: test/test-reports/python-pytest/test_testing/test_testing-509238380d0e9f60.xml (deflated 98%) 2025-12-04T09:58:05.5970571Z adding: test/test-reports/python-pytest/inductor.test_autoheuristic/inductor.test_autoheuristic-b1e93791ac32a57e.xml (deflated 28%) 2025-12-04T09:58:05.5971634Z adding: test/test-reports/python-pytest/inductor.test_cutedsl_template/inductor.test_cutedsl_template-47a7a96144df1709.xml (deflated 28%) 2025-12-04T09:58:05.6331449Z adding: test/test-reports/python-pytest/inductor.test_benchmark_fusion/inductor.test_benchmark_fusion-1a5c9a2f1b3021da.xml (deflated 94%) 2025-12-04T09:58:05.6333566Z adding: test/test-reports/python-pytest/inductor.test_remote_cache/inductor.test_remote_cache-e53c2a7e4fbe3b3f.xml (deflated 28%) 2025-12-04T09:58:05.6336334Z adding: test/test-reports/python-pytest/inductor.test_coordinate_descent_tuner/inductor.test_coordinate_descent_tuner-9ed05914eb073327.xml (deflated 29%) 2025-12-04T09:58:05.6413320Z adding: test/test-reports/python-pytest/inductor.test_inplace_padding/inductor.test_inplace_padding-ec8a12f82b4c1954.xml (deflated 99%) 2025-12-04T09:58:05.6415609Z adding: test/test-reports/python-pytest/inductor.test_cudacodecache/inductor.test_cudacodecache-c55849964bd60b32.xml (deflated 98%) 2025-12-04T09:58:05.6416673Z adding: test/test-reports/python-pytest/inductor.test_minifier_utils/inductor.test_minifier_utils-574f06f5728a4f68.xml (deflated 28%) 2025-12-04T09:58:05.6417695Z adding: test/test-reports/python-pytest/inductor.test_debug_trace/inductor.test_debug_trace-ffbc75868f825ab3.xml (deflated 28%) 2025-12-04T09:58:05.6418661Z adding: test/test-reports/python-pytest/export.test_tree_utils/export.test_tree_utils-7da9dd1b94d7565c.xml (deflated 28%) 2025-12-04T09:58:05.6419665Z adding: test/test-reports/python-pytest/inductor.test_triton_wrapper/inductor.test_triton_wrapper-0b0941e28d632943.xml (deflated 28%) 2025-12-04T09:58:05.6420736Z adding: test/test-reports/python-pytest/inductor.test_static_cuda_launcher/inductor.test_static_cuda_launcher-0c2b4708e9bf14de.xml (deflated 28%) 2025-12-04T09:58:05.6421854Z adding: test/test-reports/python-pytest/inductor.test_provenance_tracing/inductor.test_provenance_tracing-e8a06af29d1ce55a.xml (deflated 28%) 2025-12-04T09:58:05.6463441Z adding: test/test-reports/python-pytest/inductor.test_memory_planning/inductor.test_memory_planning-7b0b7b5cd4d35939.xml (deflated 98%) 2025-12-04T09:58:05.6464448Z adding: test/test-reports/python-pytest/export.test_cpp_serdes/export.test_cpp_serdes-28f96f522b58f753.xml (deflated 28%) 2025-12-04T09:58:05.6465407Z adding: test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-22c864d375691836.xml (deflated 28%) 2025-12-04T09:58:05.6466352Z adding: test/test-reports/python-pytest/test_sort_and_select/test_sort_and_select-900006a153e58acf.xml (deflated 28%) 2025-12-04T09:58:05.6467366Z adding: test/test-reports/python-pytest/functorch.test_rearrange/functorch.test_rearrange-91f40a7195b2f228.xml (deflated 28%) 2025-12-04T09:58:05.6468253Z adding: test/test-reports/python-pytest/test_package/test_package-4f9aecb7033550fa.xml (deflated 96%) 2025-12-04T09:58:05.6469053Z adding: test/test-reports/python-pytest/test_mkl_verbose/test_mkl_verbose-7e432f71a73d0aad.xml (deflated 28%) 2025-12-04T09:58:05.6470063Z adding: test/test-reports/python-pytest/test_utils_config_module/test_utils_config_module-197d280bfbc0adc1.xml (deflated 28%) 2025-12-04T09:58:05.6470942Z adding: test/test-reports/python-pytest/test_hop_infra/test_hop_infra-3a90bb4fa75fbe13.xml (deflated 28%) 2025-12-04T09:58:05.6471873Z adding: test/test-reports/python-pytest/test_appending_byte_serializer/test_appending_byte_serializer-6e708074a2279afc.xml (deflated 28%) 2025-12-04T09:58:05.6472825Z adding: test/test-reports/python-pytest/test_ao_sparsity/test_ao_sparsity-51844dfab6f78c47.xml (deflated 28%) 2025-12-04T09:58:05.6473702Z adding: test/test-reports/python-pytest/test_extension_utils/test_extension_utils-9ca2aaa70629c80b.xml (deflated 28%) 2025-12-04T09:58:05.6474607Z adding: test/test-reports/python-pytest/nn.attention.test_fa4/nn.attention.test_fa4-d1bda12ba20d2196.xml (deflated 29%) 2025-12-04T09:58:05.6475666Z adding: test/test-reports/python-pytest/typing.test_python_operators/typing.test_python_operators-ba4b7d5e3b95dbfb.xml (deflated 28%) 2025-12-04T09:58:05.6476626Z adding: test/test-reports/python-pytest/torch_np.test_dtype/torch_np.test_dtype-58897eedc88355ca.xml (deflated 28%) 2025-12-04T09:58:05.6477467Z adding: test/test-reports/python-pytest/test_file_check/test_file_check-9ff5837f5f97472d.xml (deflated 28%) 2025-12-04T09:58:05.6478335Z adding: test/test-reports/python-pytest/profiler.test_kineto/profiler.test_kineto-81a44d1732bd85bc.xml (deflated 28%) 2025-12-04T09:58:05.6479422Z adding: test/test-reports/python-pytest/functorch.test_ac_knapsack/functorch.test_ac_knapsack-56b46ef5e2d4cc8d.xml (deflated 28%) 2025-12-04T09:58:05.6480436Z adding: test/test-reports/python-pytest/torch_np.test_nep50_examples/torch_np.test_nep50_examples-a5a9215260d58c1d.xml (deflated 28%) 2025-12-04T09:58:05.6481317Z adding: test/test-reports/python-pytest/test_torch/test_torch-3d54dab073791e89.xml (deflated 97%) 2025-12-04T09:58:05.6482086Z adding: test/test-reports/python-pytest/xpu.test_gemm/xpu.test_gemm-b46323f728df1f56.xml (deflated 28%) 2025-12-04T09:58:05.6482901Z adding: test/test-reports/python-pytest/test_binary_ufuncs/test_binary_ufuncs-3281db6a17fd8e68.xml (deflated 27%) 2025-12-04T09:58:05.6483718Z adding: test/test-reports/python-pytest/test_modules/test_modules-0165bb40c77358d1.xml (deflated 98%) 2025-12-04T09:58:05.6484693Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.linalg.test_linalg/torch_np.numpy_tests.linalg.test_linalg-696665a6c90efdc4.xml (deflated 28%) 2025-12-04T09:58:05.6485852Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dtype/torch_np.numpy_tests.core.test_dtype-f968ba80b11f10a3.xml (deflated 28%) 2025-12-04T09:58:05.6486910Z adding: test/test-reports/python-pytest/lazy.test_debug_util/lazy.test_debug_util-953e7ce7fc6caa5f.xml (deflated 28%) 2025-12-04T09:58:05.6487870Z adding: test/test-reports/python-pytest/nn.test_load_state_dict/nn.test_load_state_dict-48f47027a0649f8b.xml (deflated 28%) 2025-12-04T09:58:05.6488718Z adding: test/test-reports/python-pytest/test_shape_ops/test_shape_ops-85cca46ea4083bfa.xml (deflated 28%) 2025-12-04T09:58:05.6489654Z adding: test/test-reports/python-pytest/profiler.test_memory_profiler/profiler.test_memory_profiler-b02e0fce99e01635.xml (deflated 98%) 2025-12-04T09:58:05.6490582Z adding: test/test-reports/python-pytest/test_indexing/test_indexing-c851479f46f8a70b.xml (deflated 28%) 2025-12-04T09:58:05.6491380Z adding: test/test-reports/python-pytest/test_type_info/test_type_info-1c1348d8e4ba037e.xml (deflated 28%) 2025-12-04T09:58:05.6492298Z adding: test/test-reports/python-pytest/functorch.test_aotdispatch/functorch.test_aotdispatch-0f6db4e497a4ab5e.xml (deflated 28%) 2025-12-04T09:58:05.6493297Z adding: test/test-reports/python-pytest/test_scatter_gather_ops/test_scatter_gather_ops-96cbf94077b2f70a.xml (deflated 28%) 2025-12-04T09:58:05.6494203Z adding: test/test-reports/python-pytest/test_cuda_multigpu/test_cuda_multigpu-4895f15a6c4e5281.xml (deflated 98%) 2025-12-04T09:58:05.6495288Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.lib.test_index_tricks/torch_np.numpy_tests.lib.test_index_tricks-506b926e701f49cf.xml (deflated 28%) 2025-12-04T09:58:05.6496316Z adding: test/test-reports/python-pytest/test_jit_autocast/test_jit_autocast-94bb29b8a8a5a7c3.xml (deflated 28%) 2025-12-04T09:58:05.6497320Z adding: test/test-reports/python-pytest/test_xnnpack_integration/test_xnnpack_integration-802360b65a8d44de.xml (deflated 28%) 2025-12-04T09:58:05.6498812Z adding: test/test-reports/python-pytest/nn.test_init/nn.test_init-da81cdd3c2c0ce98.xml (deflated 28%) 2025-12-04T09:58:05.6499808Z adding: test/test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-5cb9f1a76518723d.xml (deflated 99%) 2025-12-04T09:58:05.6500716Z adding: test/test-reports/python-pytest/test_type_promotion/test_type_promotion-365dfbbc94f879c8.xml (deflated 27%) 2025-12-04T09:58:05.6501668Z adding: test/test-reports/python-pytest/test_reductions/test_reductions-88d67cc570f30743.xml (deflated 28%) 2025-12-04T09:58:05.6502589Z adding: test/test-reports/python-unittest/test_autoload/TEST-TestDeviceBackendAutoload-20251204095539.xml (deflated 42%) 2025-12-04T09:58:05.6549337Z ##[group]Run # Remove any previous usage logs if they exist 2025-12-04T09:58:05.6549769Z # Remove any previous usage logs if they exist 2025-12-04T09:58:05.6550123Z rm -f logs-*.zip 2025-12-04T09:58:05.6550465Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-12-04T09:58:05.6550955Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-12-04T09:58:05.6559914Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:05.6560277Z env: 2025-12-04T09:58:05.6560489Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:05.6560743Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:05.6561060Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:05.6561590Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:05.6562052Z DEVICE_NAME: 2025-12-04T09:58:05.6562273Z DEVICE_TYPE: 2025-12-04T09:58:05.6562629Z FILE_SUFFIX: test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167 2025-12-04T09:58:05.6563041Z ##[endgroup] 2025-12-04T09:58:05.6625227Z adding: usage_log.txt (deflated 58%) 2025-12-04T09:58:05.6669121Z adding: test/test-reports/inductor.test_aot_inductor_2.5_0c57284626130d23_.log (deflated 94%) 2025-12-04T09:58:05.6670608Z adding: test/test-reports/inductor.test_torchinductor_codegen_dynamic_shapes_1.4_6c2fdc6084ab6b1c_.log (deflated 53%) 2025-12-04T09:58:05.6672174Z adding: test/test-reports/inductor.test_torchinductor_opinfo_4.14_349056dae69d6aaf_.log (deflated 96%) 2025-12-04T09:58:05.6673609Z adding: test/test-reports/inductor.test_torchinductor_opinfo_12.14_d073bcb959e81001_.log (deflated 97%) 2025-12-04T09:58:05.6675427Z adding: test/test-reports/inductor.test_flex_attention_6.6_b311f051d2033e5b_.log (deflated 96%) 2025-12-04T09:58:05.6676655Z adding: test/test-reports/inductor.test_fp8_1.1_ce1db0883598be82_.log (deflated 49%) 2025-12-04T09:58:05.6677503Z adding: test/test-reports/dynamo.test_model_output_1.1_bdadec5e5fc8d868_.log (deflated 94%) 2025-12-04T09:58:05.6678227Z adding: test/test-reports/inductor.test_triton_kernels_1.1_7dfa100f00a9f3af_.log (deflated 87%) 2025-12-04T09:58:05.6680055Z adding: test/test-reports/inductor.test_loop_ordering_1.1_2041ca8ae7154f7f_.log (deflated 95%) 2025-12-04T09:58:05.6680730Z adding: test/test-reports/export.test_serdes_1.1_6336b60dadd71d1f_.log (deflated 49%) 2025-12-04T09:58:05.6681439Z adding: test/test-reports/inductor.test_scatter_optimization_1.1_34296d74f22943a6_.log (deflated 51%) 2025-12-04T09:58:05.6682147Z adding: test/test-reports/inductor.test_padding_1.1_6c7d5d1307b4cd62_.log (deflated 49%) 2025-12-04T09:58:05.6682814Z adding: test/test-reports/dynamo.test_callback_1.1_1fd2e27dc827492e_.log (deflated 49%) 2025-12-04T09:58:05.6683644Z adding: test/test-reports/inductor.test_custom_op_autotune_1.1_d3a7b75a87bd527a_.log (deflated 51%) 2025-12-04T09:58:05.6708962Z adding: test/test-reports/test_cuda_1.1_2d01c4f5126b94de_.log (deflated 99%) 2025-12-04T09:58:05.6709550Z adding: test/test-reports/test_sparse_1.1_f9df9789c45a0117_.log (deflated 48%) 2025-12-04T09:58:05.6710183Z adding: test/test-reports/test_ci_sanity_check_fail_1.1_7w3kgi0w_toprint.log (deflated 50%) 2025-12-04T09:58:05.6710864Z adding: test/test-reports/test_ops_fwd_gradients_6.12_48d002fd493d49ca_.log (deflated 49%) 2025-12-04T09:58:05.6711502Z adding: test/test-reports/test_ops_gradients_2.10_5a1c7cb33b683fe9_.log (deflated 49%) 2025-12-04T09:58:05.6712131Z adding: test/test-reports/test_ops_gradients_10.10_499b989a7051c9dd_.log (deflated 49%) 2025-12-04T09:58:05.6712761Z adding: test/test-reports/functorch.test_ops_3.6_223700a0811fc5e7_.log (deflated 49%) 2025-12-04T09:58:05.6713504Z adding: test/test-reports/dynamo.test_after_aot_1.1_27387dde59e614f4_.log (deflated 49%) 2025-12-04T09:58:05.6714173Z adding: test/test-reports/inductor.test_snode_runtime_1.1_c167dbf323fadd03_.log (deflated 50%) 2025-12-04T09:58:05.6722884Z adding: test/test-reports/inductor.test_compiled_autograd_1.2_1749786211dd0efc_.log (deflated 98%) 2025-12-04T09:58:05.6724881Z adding: test/test-reports/test_testing_1.1_ee905ad7d74b3424_.log (deflated 96%) 2025-12-04T09:58:05.6725550Z adding: test/test-reports/inductor.test_autoheuristic_1.1_667ce2ef9a0c4c6d_.log (deflated 50%) 2025-12-04T09:58:05.6726273Z adding: test/test-reports/inductor.test_cutedsl_template_1.1_764302bf470628f6_.log (deflated 51%) 2025-12-04T09:58:05.6771817Z adding: test/test-reports/inductor.test_benchmark_fusion_1.1_baf135911d2d356d_.log (deflated 94%) 2025-12-04T09:58:05.6773259Z adding: test/test-reports/inductor.test_remote_cache_1.1_c953161b101bcc90_.log (deflated 52%) 2025-12-04T09:58:05.6774750Z adding: test/test-reports/inductor.test_coordinate_descent_tuner_1.1_d79d4fca9c00c8c8_.log (deflated 51%) 2025-12-04T09:58:05.6776269Z adding: test/test-reports/inductor.test_inplace_padding_1.1_85df0d7c6ccefed5_.log (deflated 92%) 2025-12-04T09:58:05.6777080Z adding: test/test-reports/inductor.test_cudacodecache_1.1_a3b71ead0c191355_.log (deflated 96%) 2025-12-04T09:58:05.6777795Z adding: test/test-reports/inductor.test_minifier_utils_1.1_6733dac6651991a4_.log (deflated 50%) 2025-12-04T09:58:05.6778501Z adding: test/test-reports/inductor.test_debug_trace_1.1_97586d0a33b2ca76_.log (deflated 50%) 2025-12-04T09:58:05.6779173Z adding: test/test-reports/export.test_tree_utils_1.1_f24f25a473c693f0_.log (deflated 50%) 2025-12-04T09:58:05.6779863Z adding: test/test-reports/inductor.test_triton_wrapper_1.1_e94cfd406374d935_.log (deflated 50%) 2025-12-04T09:58:05.6780605Z adding: test/test-reports/inductor.test_static_cuda_launcher_1.1_d53c30cafc20ce47_.log (deflated 51%) 2025-12-04T09:58:05.6781359Z adding: test/test-reports/inductor.test_provenance_tracing_1.1_870944fe6e309527_.log (deflated 51%) 2025-12-04T09:58:05.6811810Z adding: test/test-reports/inductor.test_memory_planning_1.1_ab660c17d0398f84_.log (deflated 98%) 2025-12-04T09:58:05.6813200Z adding: test/test-reports/export.test_cpp_serdes_1.1_5b71872ad925eef5_.log (deflated 49%) 2025-12-04T09:58:05.6814558Z adding: test/test-reports/inductor.test_control_flow_2.4_594d15ae78e9e345_.log (deflated 50%) 2025-12-04T09:58:05.6815871Z adding: test/test-reports/test_sort_and_select_1.1_1cd4117a3aecc388_.log (deflated 49%) 2025-12-04T09:58:05.6816808Z adding: test/test-reports/functorch.test_rearrange_1.1_a4effedcd0fd4055_.log (deflated 50%) 2025-12-04T09:58:05.6817464Z adding: test/test-reports/test_package_1.1_49fdbbc87a24207c_.log (deflated 92%) 2025-12-04T09:58:05.6818059Z adding: test/test-reports/test_mkl_verbose_1.1_13126a41f6f75547_.log (deflated 48%) 2025-12-04T09:58:05.6818700Z adding: test/test-reports/test_utils_config_module_1.1_619e18307e64a1f2_.log (deflated 50%) 2025-12-04T09:58:05.6819408Z adding: test/test-reports/test_hop_infra_1.1_39540bfa16752ae6_.log (deflated 48%) 2025-12-04T09:58:05.6820077Z adding: test/test-reports/test_appending_byte_serializer_1.1_5c129e37d69d7dd8_.log (deflated 50%) 2025-12-04T09:58:05.6820758Z adding: test/test-reports/test_ao_sparsity_1.1_a93db1bf7e7d6f1b_.log (deflated 48%) 2025-12-04T09:58:05.6821384Z adding: test/test-reports/test_extension_utils_1.1_82e1cd549caa8952_.log (deflated 49%) 2025-12-04T09:58:05.6822021Z adding: test/test-reports/nn.attention.test_fa4_1.1_c635564b7b730bc9_.log (deflated 49%) 2025-12-04T09:58:05.6822706Z adding: test/test-reports/typing.test_python_operators_1.1_042f5ad902ba34e2_.log (deflated 50%) 2025-12-04T09:58:05.6823380Z adding: test/test-reports/torch_np.test_dtype_1.1_26a6994970880ab6_.log (deflated 49%) 2025-12-04T09:58:05.6823987Z adding: test/test-reports/test_file_check_1.1_a601f07c53907b26_.log (deflated 48%) 2025-12-04T09:58:05.6824680Z adding: test/test-reports/profiler.test_kineto_1.1_9a415b80c30847d8_.log (deflated 49%) 2025-12-04T09:58:05.6825362Z adding: test/test-reports/functorch.test_ac_knapsack_1.1_63c0b4d8b55c8c85_.log (deflated 50%) 2025-12-04T09:58:05.6826057Z adding: test/test-reports/torch_np.test_nep50_examples_1.1_d88ffefdf604b471_.log (deflated 51%) 2025-12-04T09:58:05.6826691Z adding: test/test-reports/test_torch_1.1_a0ae95f961ed9b78_.log (deflated 95%) 2025-12-04T09:58:05.6827271Z adding: test/test-reports/xpu.test_gemm_1.1_105887c98b5f57e5_.log (deflated 48%) 2025-12-04T09:58:05.6827879Z adding: test/test-reports/test_binary_ufuncs_1.1_002f41f19aca627a_.log (deflated 49%) 2025-12-04T09:58:05.6828583Z adding: test/test-reports/test_modules_2.4_328d1f3d3b0b68db_.log (deflated 95%) 2025-12-04T09:58:05.6829267Z adding: test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_e5ceed266a7995b6_.log (deflated 52%) 2025-12-04T09:58:05.6830051Z adding: test/test-reports/torch_np.numpy_tests.core.test_dtype_1.1_12269e77e29080a2_.log (deflated 52%) 2025-12-04T09:58:05.6830760Z adding: test/test-reports/lazy.test_debug_util_1.1_ce36b8dbe1e238fe_.log (deflated 49%) 2025-12-04T09:58:05.6831398Z adding: test/test-reports/nn.test_load_state_dict_1.1_6514e2b2f6071409_.log (deflated 50%) 2025-12-04T09:58:05.6832015Z adding: test/test-reports/test_shape_ops_1.1_42da27b0199bee14_.log (deflated 48%) 2025-12-04T09:58:05.6832679Z adding: test/test-reports/profiler.test_memory_profiler_1.1_6e75fd2681e33996_.log (deflated 96%) 2025-12-04T09:58:05.6833344Z adding: test/test-reports/test_indexing_1.1_dca170216ef563f6_.log (deflated 48%) 2025-12-04T09:58:05.6833938Z adding: test/test-reports/test_type_info_1.1_586f6a1b874ca23b_.log (deflated 48%) 2025-12-04T09:58:05.6834598Z adding: test/test-reports/functorch.test_aotdispatch_1.1_e6cf91daf0dd7020_.log (deflated 52%) 2025-12-04T09:58:05.6835674Z adding: test/test-reports/test_scatter_gather_ops_1.1_7a88c857d981a5d2_.log (deflated 50%) 2025-12-04T09:58:05.6839724Z adding: test/test-reports/test_cuda_multigpu_1.1_a23ea26f9721e68d_.log (deflated 97%) 2025-12-04T09:58:05.6840457Z adding: test/test-reports/torch_np.numpy_tests.lib.test_index_tricks_1.1_529820d2d65bc59f_.log (deflated 52%) 2025-12-04T09:58:05.6841171Z adding: test/test-reports/test_jit_autocast_1.1_ec6ce096dc01ee34_.log (deflated 48%) 2025-12-04T09:58:05.6841819Z adding: test/test-reports/test_xnnpack_integration_1.1_28da4c452e743da8_.log (deflated 50%) 2025-12-04T09:58:05.6842458Z adding: test/test-reports/nn.test_init_1.1_ec9b8b7537ee12ff_.log (deflated 48%) 2025-12-04T09:58:05.6843809Z adding: test/test-reports/test_mobile_optimizer_1.1_cf6e1db294a4c552_.log (deflated 97%) 2025-12-04T09:58:05.6844459Z adding: test/test-reports/test_type_promotion_1.1_bfe6e4bccf03611c_.log (deflated 49%) 2025-12-04T09:58:05.6845087Z adding: test/test-reports/test_reductions_1.1_9d656aa654618631_.log (deflated 48%) 2025-12-04T09:58:05.6883489Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-12-04T09:58:05.6884240Z # Remove any previous debugging artifacts if they exist 2025-12-04T09:58:05.6884728Z rm -f debug-*.zip 2025-12-04T09:58:05.6885011Z if [ -d 'test/debug' ]; then 2025-12-04T09:58:05.6885359Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-12-04T09:58:05.6885681Z fi 2025-12-04T09:58:05.6894787Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:05.6895150Z env: 2025-12-04T09:58:05.6895362Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:05.6895634Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:05.6895966Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:05.6896512Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:05.6897015Z DEVICE_NAME: 2025-12-04T09:58:05.6897244Z DEVICE_TYPE: 2025-12-04T09:58:05.6897604Z FILE_SUFFIX: test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167 2025-12-04T09:58:05.6898111Z ##[endgroup] 2025-12-04T09:58:05.7007974Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T09:58:05.7008298Z with: 2025-12-04T09:58:05.7008513Z s3-bucket: gha-artifacts 2025-12-04T09:58:05.7008824Z s3-prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T09:58:05.7009153Z retention-days: 14 2025-12-04T09:58:05.7009394Z if-no-files-found: warn 2025-12-04T09:58:05.7009656Z path: test-jsons-*.zip 2025-12-04T09:58:05.7009899Z name: artifact 2025-12-04T09:58:05.7010114Z region: us-east-1 2025-12-04T09:58:05.7010329Z env: 2025-12-04T09:58:05.7010535Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:05.7010783Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:05.7011089Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:05.7011859Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:05.7012328Z DEVICE_NAME: 2025-12-04T09:58:05.7012555Z DEVICE_TYPE: 2025-12-04T09:58:05.7012770Z ##[endgroup] 2025-12-04T09:58:06.1582050Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T09:58:06.1582512Z With the provided path, there will be 1 file uploaded 2025-12-04T09:58:06.1582945Z Uploading to s3 prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T09:58:06.1657421Z Starting upload of test-jsons-test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167.zip 2025-12-04T09:58:06.3206207Z Finished upload of test-jsons-test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167.zip 2025-12-04T09:58:06.3492054Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T09:58:06.3492380Z with: 2025-12-04T09:58:06.3492609Z s3-bucket: gha-artifacts 2025-12-04T09:58:06.3492928Z s3-prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T09:58:06.3493261Z retention-days: 14 2025-12-04T09:58:06.3493530Z if-no-files-found: error 2025-12-04T09:58:06.3493807Z path: test-reports-*.zip 2025-12-04T09:58:06.3494060Z name: artifact 2025-12-04T09:58:06.3494288Z region: us-east-1 2025-12-04T09:58:06.3494509Z env: 2025-12-04T09:58:06.3494718Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:06.3494977Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:06.3495306Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:06.3495839Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:06.3496300Z DEVICE_NAME: 2025-12-04T09:58:06.3496521Z DEVICE_TYPE: 2025-12-04T09:58:06.3496741Z ##[endgroup] 2025-12-04T09:58:06.9027497Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T09:58:06.9027953Z With the provided path, there will be 1 file uploaded 2025-12-04T09:58:06.9028389Z Uploading to s3 prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T09:58:06.9102036Z Starting upload of test-reports-test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167.zip 2025-12-04T09:58:07.0817037Z Finished upload of test-reports-test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167.zip 2025-12-04T09:58:07.1110561Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T09:58:07.1110890Z with: 2025-12-04T09:58:07.1111109Z s3-bucket: gha-artifacts 2025-12-04T09:58:07.1111533Z s3-prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T09:58:07.1111861Z retention-days: 14 2025-12-04T09:58:07.1112116Z if-no-files-found: ignore 2025-12-04T09:58:07.1112378Z path: logs-*.zip 2025-12-04T09:58:07.1112602Z name: artifact 2025-12-04T09:58:07.1112824Z region: us-east-1 2025-12-04T09:58:07.1113045Z env: 2025-12-04T09:58:07.1113246Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:07.1113503Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:07.1113818Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:07.1114351Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:07.1114830Z DEVICE_NAME: 2025-12-04T09:58:07.1115050Z DEVICE_TYPE: 2025-12-04T09:58:07.1115261Z ##[endgroup] 2025-12-04T09:58:07.4431632Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T09:58:07.4432081Z With the provided path, there will be 1 file uploaded 2025-12-04T09:58:07.4432731Z Uploading to s3 prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T09:58:07.4506185Z Starting upload of logs-test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167.zip 2025-12-04T09:58:07.5933322Z Finished upload of logs-test-default-2-8-linux.g5.4xlarge.nvidia.gpu_57118183167.zip 2025-12-04T09:58:07.6236422Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T09:58:07.6236745Z with: 2025-12-04T09:58:07.6236967Z s3-bucket: gha-artifacts 2025-12-04T09:58:07.6237329Z s3-prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T09:58:07.6237665Z retention-days: 14 2025-12-04T09:58:07.6237909Z if-no-files-found: ignore 2025-12-04T09:58:07.6238176Z path: debug-*.zip 2025-12-04T09:58:07.6238593Z name: artifact 2025-12-04T09:58:07.6238819Z region: us-east-1 2025-12-04T09:58:07.6239039Z env: 2025-12-04T09:58:07.6239248Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:07.6239502Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:07.6239817Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:07.6240366Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:07.6240841Z DEVICE_NAME: 2025-12-04T09:58:07.6241062Z DEVICE_TYPE: 2025-12-04T09:58:07.6241274Z ##[endgroup] 2025-12-04T09:58:07.9455883Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-12-04T09:58:07.9749100Z ##[group]Run # shellcheck disable=SC2156 2025-12-04T09:58:07.9749469Z # shellcheck disable=SC2156 2025-12-04T09:58:07.9750028Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-12-04T09:58:07.9759336Z shell: /usr/bin/bash -e {0} 2025-12-04T09:58:07.9759593Z env: 2025-12-04T09:58:07.9759799Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:07.9760050Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:07.9760371Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:07.9760909Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:07.9761414Z DEVICE_NAME: 2025-12-04T09:58:07.9761630Z DEVICE_TYPE: 2025-12-04T09:58:07.9761850Z ##[endgroup] 2025-12-04T09:58:08.3842503Z Prepare all required actions 2025-12-04T09:58:08.3842984Z Getting action download info 2025-12-04T09:58:08.5517194Z Download action repository 'actions/setup-python@v6' (SHA:83679a892e2d95755f2dac6acb0bfd1e9ac5d548) 2025-12-04T09:58:10.0915666Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-12-04T09:58:10.0916016Z with: 2025-12-04T09:58:10.0916212Z job_id: 57118183167 2025-12-04T09:58:10.0916878Z job_name: linux-jammy-cuda12.8-py3-gcc11-slow-gradcheck / test (default, 2, 8, linux.g5.4xlarge.nvidia.gpu, module:slowgradcheck, rerun_disabled_tests) 2025-12-04T09:58:10.0917600Z workflow_name: periodic 2025-12-04T09:58:10.0917862Z workflow_run_id: 19922826259 2025-12-04T09:58:10.0918120Z workflow_attempt: 1 2025-12-04T09:58:10.0918348Z env: 2025-12-04T09:58:10.0918565Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:10.0918817Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:10.0919241Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:10.0919802Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:10.0920280Z DEVICE_NAME: 2025-12-04T09:58:10.0920504Z DEVICE_TYPE: 2025-12-04T09:58:10.0920725Z ##[endgroup] 2025-12-04T09:58:10.0981047Z ##[group]Run actions/setup-python@v6 2025-12-04T09:58:10.0981338Z with: 2025-12-04T09:58:10.0981552Z python-version: 3.10 2025-12-04T09:58:10.0981797Z check-latest: false 2025-12-04T09:58:10.0982131Z token: *** 2025-12-04T09:58:10.0982367Z update-environment: true 2025-12-04T09:58:10.0982639Z allow-prereleases: false 2025-12-04T09:58:10.0982895Z freethreaded: false 2025-12-04T09:58:10.0983126Z env: 2025-12-04T09:58:10.0983328Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:10.0983577Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:10.0983886Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:10.0984520Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:10.0984980Z DEVICE_NAME: 2025-12-04T09:58:10.0985202Z DEVICE_TYPE: 2025-12-04T09:58:10.0985413Z ##[endgroup] 2025-12-04T09:58:10.3302138Z ##[group]Installed versions 2025-12-04T09:58:10.3310992Z Version 3.10 was not found in the local cache 2025-12-04T09:58:10.3484370Z (node:102502) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. 2025-12-04T09:58:10.3485120Z (Use `node --trace-deprecation ...` to show where the warning was created) 2025-12-04T09:58:10.6797589Z ##[error]The version '3.10' with architecture 'x64' was not found for this operating system. The list of all available versions can be found here: https://raw.githubusercontent.com/actions/python-versions/main/versions-manifest.json 2025-12-04T09:58:10.7053369Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-12-04T09:58:10.7053800Z with: 2025-12-04T09:58:10.7054000Z env: 2025-12-04T09:58:10.7054215Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:10.7054483Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:10.7054797Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:10.7055326Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:10.7055804Z DEVICE_NAME: 2025-12-04T09:58:10.7056031Z DEVICE_TYPE: 2025-12-04T09:58:10.7056243Z ##[endgroup] 2025-12-04T09:58:10.7165371Z ##[group]Run set -eou pipefail 2025-12-04T09:58:10.7165677Z set -eou pipefail 2025-12-04T09:58:10.7165943Z  2025-12-04T09:58:10.7166298Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-12-04T09:58:10.7166733Z for _ in $(seq 1440); do 2025-12-04T09:58:10.7167072Z  # Break if no ssh session exists anymore 2025-12-04T09:58:10.7167482Z  if [ "$(who)" = "" ]; then 2025-12-04T09:58:10.7167764Z  break 2025-12-04T09:58:10.7167987Z  fi 2025-12-04T09:58:10.7168209Z  echo "." 2025-12-04T09:58:10.7168433Z  sleep 5 2025-12-04T09:58:10.7168662Z done 2025-12-04T09:58:10.7177643Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:10.7178037Z env: 2025-12-04T09:58:10.7178252Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:10.7178511Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:10.7178816Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:10.7179346Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:10.7179811Z DEVICE_NAME: 2025-12-04T09:58:10.7180022Z DEVICE_TYPE: 2025-12-04T09:58:10.7180236Z ##[endgroup] 2025-12-04T09:58:10.7211020Z Holding runner for 2 hours until all ssh sessions have logged out 2025-12-04T09:58:10.7659165Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:58:10.7659708Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:58:10.7660224Z # shellcheck disable=SC2046 2025-12-04T09:58:10.7660551Z docker stop $(docker ps -q) || true 2025-12-04T09:58:10.7660883Z # Prune all of the docker images 2025-12-04T09:58:10.7661193Z docker system prune -af 2025-12-04T09:58:10.7670131Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:10.7670488Z env: 2025-12-04T09:58:10.7670696Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:58:10.7670958Z HAS_NVIDIA_GPU: true 2025-12-04T09:58:10.7671272Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:58:10.7671791Z DOCKER_CONTAINER_ID: 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:10.7672261Z DEVICE_NAME: 2025-12-04T09:58:10.7672477Z DEVICE_TYPE: 2025-12-04T09:58:10.7672692Z ##[endgroup] 2025-12-04T09:58:27.6928603Z 45a59d94bb2d 2025-12-04T09:58:28.6974027Z Deleted Containers: 2025-12-04T09:58:28.6974781Z 45a59d94bb2dd1988499e18060e721f751961cba2a734c5daa30ca0af542c408 2025-12-04T09:58:28.6975108Z 2025-12-04T09:58:43.4505769Z Deleted Images: 2025-12-04T09:58:43.4506949Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:58:43.4508743Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:ba21003510dba4bdeed83df81a56fa468e0ee1b612a9445ae1f402a280804f97 2025-12-04T09:58:43.4510001Z deleted: sha256:add7313791033822205cdb3cf32096534b2cfaa4855bd48119b59000bfe00301 2025-12-04T09:58:43.4510886Z deleted: sha256:85a76b7bf29ad34eb76cce6f46af5d49a58b6272f80f983d5c769e82c7749301 2025-12-04T09:58:43.4511782Z deleted: sha256:0882f3ce59ff5ae30195ee4b059fc713e13eda107a3a7814a4616ac9058a30a4 2025-12-04T09:58:43.4513064Z deleted: sha256:64ba5b9344c11a3e4729136076830b90ac4cf1554046edb1bd4f0784b66ebd9b 2025-12-04T09:58:43.4513972Z deleted: sha256:88213c59cf461a65ab9b6cb07b4195dc9d41b5241c152daa002c7b3112e09124 2025-12-04T09:58:43.4514889Z deleted: sha256:4c0f83afa802ffbc05ebaf1aa50e48a2447c7c295549a6dded80ac63437906ca 2025-12-04T09:58:43.4515780Z deleted: sha256:6f7ec74460e8fb070c8209949095ea3be5f4e2fd69c9f750cd39ac4093f5e64b 2025-12-04T09:58:43.4516706Z deleted: sha256:d6928b0d1021b31942fdcb64e5eb4a34682de66e959dd424ed6ed02c29cd706d 2025-12-04T09:58:43.4517596Z deleted: sha256:4e9fbcb1705a6351bb34dd320558752614308636b94fd9ae6f26063e3deadc0a 2025-12-04T09:58:43.4518468Z deleted: sha256:43aabd0201f48712f21758071352dea029b4de37be08b2e2197706856a9ecbf2 2025-12-04T09:58:43.4519286Z deleted: sha256:940a98dec78303f0548beb1033242a45e9097607ef3e55c8b949b69b73d1b95e 2025-12-04T09:58:43.4520021Z deleted: sha256:d2849fa0e0411cf66e4408831d70e38838afb55b11a80c1c4d8aa0ae7dc9ca40 2025-12-04T09:58:43.4520831Z deleted: sha256:14f40d23c20c7e562623f89deb376520296758bc39dd3c77284049b84ebd8a31 2025-12-04T09:58:43.4521712Z deleted: sha256:a8ccba61f90ca097cb391d0f4fbed0d9f821d06b00e28f7332e9e2dcfcbac4ca 2025-12-04T09:58:43.4522588Z deleted: sha256:91b2060d290547d3b517d4a11d994bbe23f4560b5546cb91918ca1828dde6be1 2025-12-04T09:58:43.4523286Z deleted: sha256:b42a184755715dcfead7fad655a127433541d316d9628f5f730ff17ad5f8071c 2025-12-04T09:58:43.4523891Z deleted: sha256:aa5b4f3c9169061dc3c6da0e677e8a86f11ecb0a3f9fb4861ab3d8c04379775c 2025-12-04T09:58:43.4524506Z deleted: sha256:b4dcf450081a48d77fea0a21b8d810a69c03608a595e754fe7d365058d0579b7 2025-12-04T09:58:43.4525109Z deleted: sha256:4f7fe12d3d4f5bf890c7ada4ce16f17a105472aa6509a778f917dcce2f28174b 2025-12-04T09:58:43.4525720Z deleted: sha256:2d1d5a74182594f9a8553df00fdcfc809dba407bcd6700d667f862cbe9d555ce 2025-12-04T09:58:43.4526334Z deleted: sha256:d901e2f5d449aeed16b727bdcc11fc0e0f6c30c8fc5c39ac7eeac8a74d9d176c 2025-12-04T09:58:43.4526939Z deleted: sha256:a04df2603bd12372c6632469a9a81ebc4a8d677452c250672b9692884fa6a452 2025-12-04T09:58:43.4527636Z deleted: sha256:f438a6b52273a552dc3820a55c74c53a62a0eae9f2a7d21b37125add7d71639f 2025-12-04T09:58:43.4528243Z deleted: sha256:d4b09517e9518d709ac98b0ae6f8446ec9ac51688253607b1fca67aa2c87b3f4 2025-12-04T09:58:43.4540964Z deleted: sha256:c1fa38335237f5e7263e39d3d3de98215bcfbbb12b826955c02e149bf68efd13 2025-12-04T09:58:43.4541623Z deleted: sha256:c898d20a30de901fca74d7611663b17ab48e1726a11e031e40548ed16ee81877 2025-12-04T09:58:43.4542240Z deleted: sha256:3baceec7096518fcc10696feba551639d698b3145c2fc09cac927bb60c0fd751 2025-12-04T09:58:43.4542847Z deleted: sha256:5245aaaa3d5c3a19f76b9a6c920bd82d1a0ff5289f87c8c109652089709d9b3b 2025-12-04T09:58:43.4543440Z deleted: sha256:f05cc789b95246938c377f474c41187965b89ceac0250e7d5124bec32153f447 2025-12-04T09:58:43.4544028Z deleted: sha256:07ec4fc008de4e7a2c794ec7094cc72e0d287c04c8b2156163aee0bae147fe2d 2025-12-04T09:58:43.4544633Z deleted: sha256:c6302601ad5fde573c1f8c900250478fca7fdc6907d8fd4fae651b94b4d9264d 2025-12-04T09:58:43.4545246Z deleted: sha256:cc5e955ee1dc54931f02606c5ea87aae14f03b5d764492be611480ab041f2882 2025-12-04T09:58:43.4545848Z deleted: sha256:f21c03518996d98452338f4e80bcfd9b139a1dab155f4830be0d3f623035269f 2025-12-04T09:58:43.4546602Z deleted: sha256:519ca6f1279f7886f25f0005527cfa627deebbc5b7d7cdbfa7ef962bcfc4c26d 2025-12-04T09:58:43.4547190Z deleted: sha256:0ef990495216807d0175b192045be3f617e72331bc373b3434807f41bf69168d 2025-12-04T09:58:43.4547783Z deleted: sha256:7093edf7319e1f0e01654c3224e32c8dede5b948d106e0b9b03cbf0bb1091e33 2025-12-04T09:58:43.4548373Z deleted: sha256:c478161e058e2f4041555c3e880b95ee1ee047938dc58549a3a88135740996ae 2025-12-04T09:58:43.4548963Z deleted: sha256:9bb853b0d938cd7c36a80ce8ee40653f2c0ff92719209b11beb03acc8855ce3e 2025-12-04T09:58:43.4549577Z deleted: sha256:fdf2ace71a78ce6910ef9c4b073c195531da47022443b606bb92dcd6499b6afc 2025-12-04T09:58:43.4550180Z deleted: sha256:576c2b3770d871937d3cfb7014328bcb4bd1aed0c28bc438764b3bfdac4c1ac2 2025-12-04T09:58:43.4550774Z deleted: sha256:878e92b9cb82de09ac14a9d5f3f7bc2411a799b6f54d0d64b78c2bb4d1fdc0fc 2025-12-04T09:58:43.4551534Z deleted: sha256:85c8c3b98b65a6695f988a10cc66c981d73a3ef03eda15b8e14d227b50b56300 2025-12-04T09:58:43.4552151Z deleted: sha256:ce2ab3ba07794f9ee95d6ea7de6dcd3d2aed96561f9a79192dd56ca5bf29313a 2025-12-04T09:58:43.4552751Z deleted: sha256:37a6e12976ca957286977e696e63012ab9821214b0483fe1a48d29dcb280508a 2025-12-04T09:58:43.4553339Z deleted: sha256:cd1d5d3dd7038144ca6fe961c0d4c8e705625ae0c36190ba8b3e9602abedad19 2025-12-04T09:58:43.4553934Z deleted: sha256:0e707276e0be2e0008b86d594fadc0d16444d66c4fb7227c56f144cbb3c2affd 2025-12-04T09:58:43.4554532Z deleted: sha256:22d4aad6a2ada91b341c1225a0f314042b8aeabef7568c5c019709b058bf070b 2025-12-04T09:58:43.4555134Z deleted: sha256:ee4adacf4e0933131d0275eddad406b3c8147e6cf07a292b99f1aff4b5355f33 2025-12-04T09:58:43.4555739Z deleted: sha256:43da0b9e7c0e18403dcb834e53628dc7c970ccb2dbd091878c0d7c0170dbc97f 2025-12-04T09:58:43.4556341Z deleted: sha256:00571684bdcd75beda15eb7d4e79b5458bc914350f9bb4d87fcdc97ad15e0da1 2025-12-04T09:58:43.4556945Z deleted: sha256:41615f09950259f1d75e82ef35b6fc53b18fe71ebff143744cfd51009d04349e 2025-12-04T09:58:43.4557547Z deleted: sha256:75ab34d2eed3c7915467a506ab6dab2711918fbabe94add2fb5c62780221ab0c 2025-12-04T09:58:43.4558161Z deleted: sha256:0a39ef2bebf44c1c3893d1e5fb42dad48b8fac7ca673141267ee967f85455e89 2025-12-04T09:58:43.4558763Z deleted: sha256:9b7d024e48ba1f9824a54597621b1b062cbc4aa41a77d81ca538d6b5c24a612c 2025-12-04T09:58:43.4559339Z deleted: sha256:392257172de6434c271bd93394218a91e9aa86d7c18abc2f2759317b9d5fb6de 2025-12-04T09:58:43.4559911Z deleted: sha256:6c3232860b930866a463a356124fc392c7e5f04895695229257e8c3e8a02711d 2025-12-04T09:58:43.4560491Z deleted: sha256:63dd55b807215e2fa6c715419ac0c5072d02dddc848dbf74bb7e77b906b5eaed 2025-12-04T09:58:43.4561081Z deleted: sha256:07a8738c1b4584db72ed9aa60f5274321eb0ba16263450da3a75df8326ebc25f 2025-12-04T09:58:43.4561659Z deleted: sha256:053fe2965b01281d12040ec1893e0d1aa77362a49ea9a1067402272c69dad9f5 2025-12-04T09:58:43.4562250Z deleted: sha256:7857fb5eb181c4e80262ecab60bdd3c266cf3d1409ceb76c05882609b416a8d3 2025-12-04T09:58:43.4562856Z deleted: sha256:752528477fc99089de3bd2c6da7b30cf34f2e901fe06d8fcfe685b411461e883 2025-12-04T09:58:43.4563530Z deleted: sha256:cce0210e2f4b042601813df03aa294a86b0c668fcfc75f4c63f6fa12b2952e15 2025-12-04T09:58:43.4564123Z deleted: sha256:f2bb405a26705ecd12d21380d26d9355d01db3a2175080fbdb468f2b5a25a76c 2025-12-04T09:58:43.4564727Z deleted: sha256:ad430120d4ffbaf97cd8d6de6ea8eefa4a8f80ec45f0b176c6b26bff0970fd33 2025-12-04T09:58:43.4565340Z deleted: sha256:225a4910baea7cc540ed43eeac75046293800ab0b8e0192b51e991c8cb50bcf3 2025-12-04T09:58:43.4565932Z deleted: sha256:a259945b0c3507f049fbac10fb3d3ffe43d45e83c91b80ae8cd1dafb855ad83c 2025-12-04T09:58:43.4566538Z deleted: sha256:862a98881b1d5adad5c21d01602773b894794097de80964ef8f47bcaadb43255 2025-12-04T09:58:43.4567280Z deleted: sha256:1cf6d3c8b6c2694b79a2d08719594903811c330a36a4c7a8a7153a350b53d292 2025-12-04T09:58:43.4567879Z deleted: sha256:232a1ae8b0fee817ff7838bb5986a2f38377d3b1dbbf5217b576af0f953b0844 2025-12-04T09:58:43.4568484Z deleted: sha256:c72c5705dabd6314423dd7d4fb260a20d5d9886b2ebce60d19e9d78c4a2335c2 2025-12-04T09:58:43.4569134Z deleted: sha256:296734cf81fd92c913884d058908598424ffe072676e38de289bbab83768c7bd 2025-12-04T09:58:43.4569720Z deleted: sha256:7c76040481b889847a1804021aeff07547eaa4ee706d6137db218d497a8fd9c1 2025-12-04T09:58:43.4570310Z deleted: sha256:d5e293f5b354e8cbcc6de893ea72cc632b02d8fdfbb08ec3127c4e9662f3ebff 2025-12-04T09:58:43.4570916Z deleted: sha256:f35a64e429c88e249645090f21fbe7dae108d98e0ab4ea13184f24b3fd66c315 2025-12-04T09:58:43.4571515Z deleted: sha256:ce6ae8d595c8e69115c51b1ce4f9a9158795d7b863b1cb53f21c39a87974d41b 2025-12-04T09:58:43.4572119Z deleted: sha256:8941abaee59400fb9b3a60765fea4a1fc2a6a447467a6d983e84c7f72494a450 2025-12-04T09:58:43.4572719Z deleted: sha256:ef53c29a9a2c2bc80ffdb9bfaf92842436b5755ec1ce828b9d11e5e27d656ea1 2025-12-04T09:58:43.4573328Z deleted: sha256:7a347fb0acb43f1c814f8c8ff21185e8b5cf64d7bc5988cea060f77d906e08b5 2025-12-04T09:58:43.4574019Z deleted: sha256:cc855dc9be79496e15175569dced2d13477e50b077a5fd3945f9bf50018880c1 2025-12-04T09:58:43.4574615Z deleted: sha256:f7a9946ada3d4786658bc0b643808bb32a9a45e4e90e30dc43ee19e2dbe24024 2025-12-04T09:58:43.4575213Z deleted: sha256:c22a9215f62812c1d2e32827f5221ff556c5b6702aadbdab6b87b8293f19635e 2025-12-04T09:58:43.4575805Z deleted: sha256:959a56746620012e37c1def1a83c5afb1e7c0adc59b021a28beb53c24df98032 2025-12-04T09:58:43.4576459Z deleted: sha256:31a0fff0695bf6100c17954be72eab2095b466d559c75c3faf2a17d8c41e6ebe 2025-12-04T09:58:43.4577051Z deleted: sha256:c15e2b5241b9e55af1b2593e544391b4b44d0505e6528e8f12425136e93b424c 2025-12-04T09:58:43.4577640Z deleted: sha256:73974f74b436f39a2fdb6461b1e3f7c3e41c73325776fa71d16b942a5b4a365b 2025-12-04T09:58:43.4578139Z untagged: public.ecr.aws/docker/library/python:3.13 2025-12-04T09:58:43.4578800Z untagged: public.ecr.aws/docker/library/python@sha256:3f986299a7b8b44b0d8cf9bda2b22361ce5c3058ef5d7cb17fb7452506680ab0 2025-12-04T09:58:43.4579587Z deleted: sha256:44438aecfedf7b6086fce506dae0db5ba7fc0027f9b743f1a75a6b5cbc7de70a 2025-12-04T09:58:43.4580208Z deleted: sha256:6f09a1f5d8a107c2532fbd116e75116cb75fa77b1a7d72d3bdf1ac12de152acd 2025-12-04T09:58:43.4580820Z deleted: sha256:fe5f3ac0be086125eb1e3cd10cc33e8e426f4e079381f7ce5a987b626e99fa67 2025-12-04T09:58:43.4581423Z deleted: sha256:79dd2061a22cf919cfc4f1f02704bfda09afadb017265e670ee54441d296c06c 2025-12-04T09:58:43.4582042Z deleted: sha256:9447ad402aafdbee17e999b0ec84ad89c2646dbebf054d469d4f8bee77f66212 2025-12-04T09:58:43.4582639Z deleted: sha256:7a4909f3c1975be52292f53107495ee1b41c17494918767ccedf1cf1688ae318 2025-12-04T09:58:43.4583210Z deleted: sha256:3474923d97f1f498237650a7d51bd4aea37d5e6b9d8a778777920584af5dd560 2025-12-04T09:58:43.4583806Z deleted: sha256:683afd1773444401a9cbd24842ee5d9154a11abb4fab63ddea5c03df788597ee 2025-12-04T09:58:43.4584164Z 2025-12-04T09:58:43.4584287Z Total reclaimed space: 35.64GB 2025-12-04T09:58:43.4659040Z Post job cleanup. 2025-12-04T09:58:43.4694739Z Post job cleanup. 2025-12-04T09:58:43.6046938Z (node:102607) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. 2025-12-04T09:58:43.6048112Z (Use `node --trace-deprecation ...` to show where the warning was created) 2025-12-04T09:58:43.6240160Z Post job cleanup. 2025-12-04T09:58:43.6297782Z Post job cleanup. 2025-12-04T09:58:43.7272681Z [command]/usr/bin/git version 2025-12-04T09:58:43.7313843Z git version 2.50.1 2025-12-04T09:58:43.7349559Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/82a7ae02-e33e-4161-8ea7-6820b5259fa5/.gitconfig' 2025-12-04T09:58:43.7360587Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/82a7ae02-e33e-4161-8ea7-6820b5259fa5' before making global git config changes 2025-12-04T09:58:43.7361494Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T09:58:43.7366404Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:58:43.7413722Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T09:58:43.7462391Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T09:58:43.7870790Z Entering 'android/libs/fbjni' 2025-12-04T09:58:43.7961234Z Entering 'third_party/FP16' 2025-12-04T09:58:43.8040701Z Entering 'third_party/FXdiv' 2025-12-04T09:58:43.8121808Z Entering 'third_party/NNPACK' 2025-12-04T09:58:43.8209520Z Entering 'third_party/NVTX' 2025-12-04T09:58:43.8290705Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:58:43.8372074Z Entering 'third_party/XNNPACK' 2025-12-04T09:58:43.8473510Z Entering 'third_party/aiter' 2025-12-04T09:58:43.8554881Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:58:43.8644143Z Entering 'third_party/benchmark' 2025-12-04T09:58:43.8723867Z Entering 'third_party/composable_kernel' 2025-12-04T09:58:43.8814100Z Entering 'third_party/cpp-httplib' 2025-12-04T09:58:43.8894597Z Entering 'third_party/cpuinfo' 2025-12-04T09:58:43.8977301Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:58:43.9061082Z Entering 'third_party/cutlass' 2025-12-04T09:58:43.9152235Z Entering 'third_party/fbgemm' 2025-12-04T09:58:43.9234513Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:58:43.9312561Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:58:43.9399892Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:58:43.9477995Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:58:43.9570065Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:58:43.9649554Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:58:43.9725616Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:58:43.9810077Z Entering 'third_party/flash-attention' 2025-12-04T09:58:43.9893320Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:58:43.9977406Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:58:44.0066662Z Entering 'third_party/flatbuffers' 2025-12-04T09:58:44.0152087Z Entering 'third_party/fmt' 2025-12-04T09:58:44.0232525Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:58:44.0314748Z Entering 'third_party/gloo' 2025-12-04T09:58:44.0395849Z Entering 'third_party/googletest' 2025-12-04T09:58:44.0476781Z Entering 'third_party/ideep' 2025-12-04T09:58:44.0554655Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:58:44.0644143Z Entering 'third_party/ittapi' 2025-12-04T09:58:44.0724756Z Entering 'third_party/kineto' 2025-12-04T09:58:44.0810624Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:58:44.0886933Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:58:44.0965942Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:58:44.1044701Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:58:44.1123363Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:58:44.1201367Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:58:44.1283177Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:58:44.1366236Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:58:44.1449208Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:58:44.1529974Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:58:44.1608444Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:58:44.1685400Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:58:44.1764490Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:58:44.1852509Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:58:44.1930393Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:58:44.2014616Z Entering 'third_party/kleidiai' 2025-12-04T09:58:44.2095605Z Entering 'third_party/mimalloc' 2025-12-04T09:58:44.2177713Z Entering 'third_party/nlohmann' 2025-12-04T09:58:44.2261671Z Entering 'third_party/onnx' 2025-12-04T09:58:44.2362641Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:58:44.2452659Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:58:44.2533746Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:58:44.2614193Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:58:44.2690770Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:58:44.2767887Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:58:44.2848779Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:58:44.2925290Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:58:44.3006254Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:58:44.3089409Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:58:44.3169263Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:58:44.3255309Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:58:44.3356107Z Entering 'third_party/pocketfft' 2025-12-04T09:58:44.3436465Z Entering 'third_party/protobuf' 2025-12-04T09:58:44.3518839Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:58:44.3597425Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:58:44.3680967Z Entering 'third_party/psimd' 2025-12-04T09:58:44.3765693Z Entering 'third_party/pthreadpool' 2025-12-04T09:58:44.3848174Z Entering 'third_party/pybind11' 2025-12-04T09:58:44.3928000Z Entering 'third_party/python-peachpy' 2025-12-04T09:58:44.4008372Z Entering 'third_party/sleef' 2025-12-04T09:58:44.4090904Z Entering 'third_party/tensorpipe' 2025-12-04T09:58:44.4170721Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:58:44.4248096Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:58:44.4323685Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:58:44.4401587Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:58:44.4481545Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:58:44.4599770Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T09:58:44.4629102Z http.https://github.com/.extraheader 2025-12-04T09:58:44.4641744Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-12-04T09:58:44.4680660Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T09:58:44.5100486Z Entering 'android/libs/fbjni' 2025-12-04T09:58:44.5154839Z http.https://github.com/.extraheader 2025-12-04T09:58:44.5207599Z Entering 'third_party/FP16' 2025-12-04T09:58:44.5264274Z http.https://github.com/.extraheader 2025-12-04T09:58:44.5314614Z Entering 'third_party/FXdiv' 2025-12-04T09:58:44.5365941Z http.https://github.com/.extraheader 2025-12-04T09:58:44.5419349Z Entering 'third_party/NNPACK' 2025-12-04T09:58:44.5471326Z http.https://github.com/.extraheader 2025-12-04T09:58:44.5522263Z Entering 'third_party/NVTX' 2025-12-04T09:58:44.5573096Z http.https://github.com/.extraheader 2025-12-04T09:58:44.5624426Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:58:44.5675358Z http.https://github.com/.extraheader 2025-12-04T09:58:44.5724606Z Entering 'third_party/XNNPACK' 2025-12-04T09:58:44.5778533Z http.https://github.com/.extraheader 2025-12-04T09:58:44.5846509Z Entering 'third_party/aiter' 2025-12-04T09:58:44.5900996Z http.https://github.com/.extraheader 2025-12-04T09:58:44.5949741Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:58:44.5999781Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6060482Z Entering 'third_party/benchmark' 2025-12-04T09:58:44.6112106Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6162669Z Entering 'third_party/composable_kernel' 2025-12-04T09:58:44.6217526Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6279207Z Entering 'third_party/cpp-httplib' 2025-12-04T09:58:44.6330748Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6384582Z Entering 'third_party/cpuinfo' 2025-12-04T09:58:44.6437101Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6485848Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:58:44.6543761Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6594650Z Entering 'third_party/cutlass' 2025-12-04T09:58:44.6646740Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6707100Z Entering 'third_party/fbgemm' 2025-12-04T09:58:44.6759463Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6810391Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:58:44.6866258Z http.https://github.com/.extraheader 2025-12-04T09:58:44.6915198Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:58:44.6965502Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7023064Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:58:44.7074604Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7125337Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:58:44.7176415Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7234870Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:58:44.7285604Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7337182Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:58:44.7386782Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7437897Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:58:44.7496458Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7551820Z Entering 'third_party/flash-attention' 2025-12-04T09:58:44.7602629Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7651195Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:58:44.7701148Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7756786Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:58:44.7805966Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7872044Z Entering 'third_party/flatbuffers' 2025-12-04T09:58:44.7923815Z http.https://github.com/.extraheader 2025-12-04T09:58:44.7977694Z Entering 'third_party/fmt' 2025-12-04T09:58:44.8029966Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8079793Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:58:44.8135020Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8186904Z Entering 'third_party/gloo' 2025-12-04T09:58:44.8239631Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8288713Z Entering 'third_party/googletest' 2025-12-04T09:58:44.8340042Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8390407Z Entering 'third_party/ideep' 2025-12-04T09:58:44.8441321Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8489624Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:58:44.8543292Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8603837Z Entering 'third_party/ittapi' 2025-12-04T09:58:44.8656689Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8709838Z Entering 'third_party/kineto' 2025-12-04T09:58:44.8762702Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8812989Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:58:44.8863481Z http.https://github.com/.extraheader 2025-12-04T09:58:44.8911940Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:58:44.8962797Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9015571Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:58:44.9066476Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9116024Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:58:44.9172349Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9226161Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:58:44.9277842Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9326135Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:58:44.9378168Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9433681Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:58:44.9484690Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9534467Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:58:44.9590273Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9643282Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:58:44.9695161Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9746856Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:58:44.9798954Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9850299Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:58:44.9900413Z http.https://github.com/.extraheader 2025-12-04T09:58:44.9949359Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:58:45.0000771Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0054799Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:58:45.0105116Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0169304Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:58:45.0219806Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0269796Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:58:45.0321275Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0374555Z Entering 'third_party/kleidiai' 2025-12-04T09:58:45.0425742Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0476077Z Entering 'third_party/mimalloc' 2025-12-04T09:58:45.0528677Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0582047Z Entering 'third_party/nlohmann' 2025-12-04T09:58:45.0635567Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0688254Z Entering 'third_party/onnx' 2025-12-04T09:58:45.0742392Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0809538Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:58:45.0864262Z http.https://github.com/.extraheader 2025-12-04T09:58:45.0921920Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:58:45.0976708Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1025708Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:58:45.1077730Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1126781Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:58:45.1180241Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1229939Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:58:45.1278679Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1327233Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:58:45.1376627Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1426819Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:58:45.1479217Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1529364Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:58:45.1579633Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1628019Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:58:45.1678615Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1725517Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:58:45.1780928Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1832937Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:58:45.1883318Z http.https://github.com/.extraheader 2025-12-04T09:58:45.1937661Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:58:45.1989816Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2063057Z Entering 'third_party/pocketfft' 2025-12-04T09:58:45.2115066Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2164200Z Entering 'third_party/protobuf' 2025-12-04T09:58:45.2216260Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2269768Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:58:45.2320466Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2369857Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:58:45.2418285Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2472643Z Entering 'third_party/psimd' 2025-12-04T09:58:45.2524710Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2573976Z Entering 'third_party/pthreadpool' 2025-12-04T09:58:45.2625814Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2675344Z Entering 'third_party/pybind11' 2025-12-04T09:58:45.2731585Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2781501Z Entering 'third_party/python-peachpy' 2025-12-04T09:58:45.2832724Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2883695Z Entering 'third_party/sleef' 2025-12-04T09:58:45.2934671Z http.https://github.com/.extraheader 2025-12-04T09:58:45.2987136Z Entering 'third_party/tensorpipe' 2025-12-04T09:58:45.3039135Z http.https://github.com/.extraheader 2025-12-04T09:58:45.3085752Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:58:45.3140862Z http.https://github.com/.extraheader 2025-12-04T09:58:45.3189404Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:58:45.3244118Z http.https://github.com/.extraheader 2025-12-04T09:58:45.3293122Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:58:45.3346870Z http.https://github.com/.extraheader 2025-12-04T09:58:45.3395063Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:58:45.3444730Z http.https://github.com/.extraheader 2025-12-04T09:58:45.3490668Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:58:45.3541708Z http.https://github.com/.extraheader 2025-12-04T09:58:45.3621726Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.3671435Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T09:58:45.4074225Z Entering 'android/libs/fbjni' 2025-12-04T09:58:45.4111340Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T09:58:45.4137142Z Entering 'third_party/FP16' 2025-12-04T09:58:45.4173339Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T09:58:45.4196169Z Entering 'third_party/FXdiv' 2025-12-04T09:58:45.4231955Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T09:58:45.4257004Z Entering 'third_party/NNPACK' 2025-12-04T09:58:45.4291496Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T09:58:45.4315318Z Entering 'third_party/NVTX' 2025-12-04T09:58:45.4349123Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T09:58:45.4375637Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:58:45.4409700Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T09:58:45.4435179Z Entering 'third_party/XNNPACK' 2025-12-04T09:58:45.4470823Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T09:58:45.4513565Z Entering 'third_party/aiter' 2025-12-04T09:58:45.4550402Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T09:58:45.4577103Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:58:45.4610068Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T09:58:45.4646233Z Entering 'third_party/benchmark' 2025-12-04T09:58:45.4680420Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:58:45.4703942Z Entering 'third_party/composable_kernel' 2025-12-04T09:58:45.4738108Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T09:58:45.4773348Z Entering 'third_party/cpp-httplib' 2025-12-04T09:58:45.4808828Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T09:58:45.4832065Z Entering 'third_party/cpuinfo' 2025-12-04T09:58:45.4865782Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T09:58:45.4892432Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:58:45.4926487Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T09:58:45.4957164Z Entering 'third_party/cutlass' 2025-12-04T09:58:45.4991406Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T09:58:45.5027963Z Entering 'third_party/fbgemm' 2025-12-04T09:58:45.5066279Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T09:58:45.5097878Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:58:45.5130320Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T09:58:45.5154504Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:58:45.5188557Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T09:58:45.5220178Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:58:45.5253992Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T09:58:45.5278279Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:58:45.5311668Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T09:58:45.5345765Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:58:45.5378929Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T09:58:45.5402416Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:58:45.5435474Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T09:58:45.5460401Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:58:45.5493443Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T09:58:45.5523532Z Entering 'third_party/flash-attention' 2025-12-04T09:58:45.5558275Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T09:58:45.5584433Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:58:45.5617763Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T09:58:45.5648793Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:58:45.5682842Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T09:58:45.5718189Z Entering 'third_party/flatbuffers' 2025-12-04T09:58:45.5753426Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T09:58:45.5782084Z Entering 'third_party/fmt' 2025-12-04T09:58:45.5820021Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:58:45.5846398Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:58:45.5881181Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T09:58:45.5908214Z Entering 'third_party/gloo' 2025-12-04T09:58:45.5944077Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T09:58:45.5969749Z Entering 'third_party/googletest' 2025-12-04T09:58:45.6003943Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:58:45.6029737Z Entering 'third_party/ideep' 2025-12-04T09:58:45.6067120Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T09:58:45.6091886Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:58:45.6125547Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T09:58:45.6160258Z Entering 'third_party/ittapi' 2025-12-04T09:58:45.6198866Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T09:58:45.6226271Z Entering 'third_party/kineto' 2025-12-04T09:58:45.6262938Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T09:58:45.6286438Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:58:45.6320243Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T09:58:45.6344079Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:58:45.6376880Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T09:58:45.6403260Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:58:45.6439794Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T09:58:45.6464104Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:58:45.6497559Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:58:45.6521623Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:58:45.6556341Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T09:58:45.6578635Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:58:45.6613445Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T09:58:45.6641660Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:58:45.6677092Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T09:58:45.6702927Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:58:45.6738802Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:58:45.6762852Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:58:45.6797695Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T09:58:45.6825770Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:58:45.6860013Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T09:58:45.6884610Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:58:45.6919945Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:58:45.6941293Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:58:45.6976285Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:58:45.7003352Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:58:45.7040050Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:58:45.7071975Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:58:45.7105287Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T09:58:45.7129656Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:58:45.7166140Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T09:58:45.7198759Z Entering 'third_party/kleidiai' 2025-12-04T09:58:45.7234741Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T09:58:45.7260296Z Entering 'third_party/mimalloc' 2025-12-04T09:58:45.7296303Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T09:58:45.7321979Z Entering 'third_party/nlohmann' 2025-12-04T09:58:45.7357198Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T09:58:45.7385370Z Entering 'third_party/onnx' 2025-12-04T09:58:45.7420294Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T09:58:45.7462909Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:58:45.7496680Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:58:45.7527672Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:58:45.7565771Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T09:58:45.7590814Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:58:45.7622728Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:58:45.7647308Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:58:45.7680571Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:58:45.7705275Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:58:45.7739539Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T09:58:45.7763058Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:58:45.7800006Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T09:58:45.7824316Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:58:45.7857377Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T09:58:45.7880813Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:58:45.7914367Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T09:58:45.7943455Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:58:45.7976475Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:58:45.7998879Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:58:45.8033274Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:58:45.8059529Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:58:45.8092552Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:58:45.8120450Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:58:45.8158880Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T09:58:45.8207908Z Entering 'third_party/pocketfft' 2025-12-04T09:58:45.8244082Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T09:58:45.8270022Z Entering 'third_party/protobuf' 2025-12-04T09:58:45.8307555Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T09:58:45.8336698Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:58:45.8371266Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:58:45.8394512Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:58:45.8428510Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:58:45.8457491Z Entering 'third_party/psimd' 2025-12-04T09:58:45.8492178Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T09:58:45.8517099Z Entering 'third_party/pthreadpool' 2025-12-04T09:58:45.8552549Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T09:58:45.8577695Z Entering 'third_party/pybind11' 2025-12-04T09:58:45.8612815Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:58:45.8638119Z Entering 'third_party/python-peachpy' 2025-12-04T09:58:45.8672626Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T09:58:45.8697658Z Entering 'third_party/sleef' 2025-12-04T09:58:45.8736762Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T09:58:45.8761891Z Entering 'third_party/tensorpipe' 2025-12-04T09:58:45.8796780Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T09:58:45.8823643Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:58:45.8858573Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:58:45.8883286Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:58:45.8918203Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T09:58:45.8941899Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:58:45.8974526Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T09:58:45.8998799Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:58:45.9033982Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:58:45.9055803Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:58:45.9089884Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T09:58:45.9150143Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9186767Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9221050Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9256263Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9290977Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9325477Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9361571Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9394877Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9429558Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9465536Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9499239Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9533870Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9568564Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9602393Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9637621Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9671779Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9705536Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9740027Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9774046Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9808235Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9840953Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9880297Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9914969Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9949046Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:45.9982798Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0016169Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0051124Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0084923Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0119458Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0152953Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0186114Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0218896Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0252857Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0287413Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0322249Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0354650Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0387348Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0421589Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0456194Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0491785Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0526369Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0561237Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0607600Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0645295Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0681192Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0718796Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0753892Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0788345Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0824001Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0859146Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0893434Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0927382Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0962575Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.0995018Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1029082Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1063796Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1098092Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1133148Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1166290Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1200271Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1236766Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1271113Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1304748Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1336447Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1377223Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1414488Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1450351Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1485557Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1533233Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1560644Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1595881Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1630022Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1665360Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1700423Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1736017Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1770430Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1804765Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1840318Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1873261Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1909062Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.1946771Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:58:46.2091452Z A job completed hook has been configured by the self-hosted runner administrator 2025-12-04T09:58:46.2112728Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-12-04T09:58:46.2120850Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:58:46.2121222Z ##[endgroup] 2025-12-04T09:58:46.2232512Z [!ALERT!] Swap in detected! [!ALERT!] 2025-12-04T09:58:57.3029140Z [!ALERT!] Swap out detected [!ALERT!] 2025-12-04T09:59:16.3647847Z Cleaning up orphan processes